00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-main" build number 3900 00:00:00.001 originally caused by: 00:00:00.002 Started by upstream project "nightly-trigger" build number 3495 00:00:00.002 originally caused by: 00:00:00.002 Started by timer 00:00:00.055 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.055 The recommended git tool is: git 00:00:00.055 using credential 00000000-0000-0000-0000-000000000002 00:00:00.057 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.082 Fetching changes from the remote Git repository 00:00:00.084 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.131 Using shallow fetch with depth 1 00:00:00.131 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.131 > git --version # timeout=10 00:00:00.192 > git --version # 'git version 2.39.2' 00:00:00.192 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.212 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.212 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:03.007 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:03.018 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:03.032 Checking out Revision 53a1a621557260e3fbfd1fd32ee65ff11a804d5b (FETCH_HEAD) 00:00:03.032 > git config core.sparsecheckout # timeout=10 00:00:03.045 > git read-tree -mu HEAD # timeout=10 00:00:03.062 > git checkout -f 53a1a621557260e3fbfd1fd32ee65ff11a804d5b # timeout=5 00:00:03.084 Commit message: "packer: Merge irdmafedora into main fedora image" 00:00:03.084 > git rev-list --no-walk 53a1a621557260e3fbfd1fd32ee65ff11a804d5b # timeout=10 00:00:03.314 [Pipeline] Start of Pipeline 00:00:03.329 [Pipeline] library 00:00:03.331 Loading library shm_lib@master 00:00:03.331 Library shm_lib@master is cached. Copying from home. 00:00:03.346 [Pipeline] node 00:00:03.358 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:03.360 [Pipeline] { 00:00:03.370 [Pipeline] catchError 00:00:03.371 [Pipeline] { 00:00:03.384 [Pipeline] wrap 00:00:03.393 [Pipeline] { 00:00:03.401 [Pipeline] stage 00:00:03.403 [Pipeline] { (Prologue) 00:00:03.424 [Pipeline] echo 00:00:03.425 Node: VM-host-WFP7 00:00:03.432 [Pipeline] cleanWs 00:00:03.442 [WS-CLEANUP] Deleting project workspace... 00:00:03.442 [WS-CLEANUP] Deferred wipeout is used... 00:00:03.449 [WS-CLEANUP] done 00:00:03.685 [Pipeline] setCustomBuildProperty 00:00:03.791 [Pipeline] httpRequest 00:00:04.199 [Pipeline] echo 00:00:04.201 Sorcerer 10.211.164.101 is alive 00:00:04.211 [Pipeline] retry 00:00:04.214 [Pipeline] { 00:00:04.227 [Pipeline] httpRequest 00:00:04.232 HttpMethod: GET 00:00:04.233 URL: http://10.211.164.101/packages/jbp_53a1a621557260e3fbfd1fd32ee65ff11a804d5b.tar.gz 00:00:04.233 Sending request to url: http://10.211.164.101/packages/jbp_53a1a621557260e3fbfd1fd32ee65ff11a804d5b.tar.gz 00:00:04.234 Response Code: HTTP/1.1 200 OK 00:00:04.235 Success: Status code 200 is in the accepted range: 200,404 00:00:04.235 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_53a1a621557260e3fbfd1fd32ee65ff11a804d5b.tar.gz 00:00:04.381 [Pipeline] } 00:00:04.391 [Pipeline] // retry 00:00:04.396 [Pipeline] sh 00:00:04.676 + tar --no-same-owner -xf jbp_53a1a621557260e3fbfd1fd32ee65ff11a804d5b.tar.gz 00:00:04.690 [Pipeline] httpRequest 00:00:05.044 [Pipeline] echo 00:00:05.045 Sorcerer 10.211.164.101 is alive 00:00:05.052 [Pipeline] retry 00:00:05.054 [Pipeline] { 00:00:05.065 [Pipeline] httpRequest 00:00:05.070 HttpMethod: GET 00:00:05.070 URL: http://10.211.164.101/packages/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:05.070 Sending request to url: http://10.211.164.101/packages/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:05.071 Response Code: HTTP/1.1 200 OK 00:00:05.072 Success: Status code 200 is in the accepted range: 200,404 00:00:05.072 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:23.474 [Pipeline] } 00:00:23.492 [Pipeline] // retry 00:00:23.501 [Pipeline] sh 00:00:23.788 + tar --no-same-owner -xf spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:26.341 [Pipeline] sh 00:00:26.626 + git -C spdk log --oneline -n5 00:00:26.626 09cc66129 test/unit: add mixed busy/idle mock poller function in reactor_ut 00:00:26.626 a67b3561a dpdk: update submodule to include alarm_cancel fix 00:00:26.626 43f6d3385 nvmf: remove use of STAILQ for last_wqe events 00:00:26.626 9645421c5 nvmf: rename nvmf_rdma_qpair_process_ibv_event() 00:00:26.626 e6da32ee1 nvmf: rename nvmf_rdma_send_qpair_async_event() 00:00:26.647 [Pipeline] withCredentials 00:00:26.658 > git --version # timeout=10 00:00:26.672 > git --version # 'git version 2.39.2' 00:00:26.691 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:26.693 [Pipeline] { 00:00:26.701 [Pipeline] retry 00:00:26.703 [Pipeline] { 00:00:26.718 [Pipeline] sh 00:00:27.003 + git ls-remote http://dpdk.org/git/dpdk main 00:00:27.277 [Pipeline] } 00:00:27.300 [Pipeline] // retry 00:00:27.305 [Pipeline] } 00:00:27.323 [Pipeline] // withCredentials 00:00:27.336 [Pipeline] httpRequest 00:00:27.770 [Pipeline] echo 00:00:27.772 Sorcerer 10.211.164.101 is alive 00:00:27.784 [Pipeline] retry 00:00:27.786 [Pipeline] { 00:00:27.803 [Pipeline] httpRequest 00:00:27.809 HttpMethod: GET 00:00:27.809 URL: http://10.211.164.101/packages/dpdk_bf0ff8df59c7e32f95c0b542cc4a7918f8a3da84.tar.gz 00:00:27.810 Sending request to url: http://10.211.164.101/packages/dpdk_bf0ff8df59c7e32f95c0b542cc4a7918f8a3da84.tar.gz 00:00:27.829 Response Code: HTTP/1.1 200 OK 00:00:27.830 Success: Status code 200 is in the accepted range: 200,404 00:00:27.830 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_bf0ff8df59c7e32f95c0b542cc4a7918f8a3da84.tar.gz 00:01:03.066 [Pipeline] } 00:01:03.083 [Pipeline] // retry 00:01:03.092 [Pipeline] sh 00:01:03.373 + tar --no-same-owner -xf dpdk_bf0ff8df59c7e32f95c0b542cc4a7918f8a3da84.tar.gz 00:01:04.764 [Pipeline] sh 00:01:05.052 + git -C dpdk log --oneline -n5 00:01:05.052 bf0ff8df59 maintainers: fix prog guide paths 00:01:05.052 41dd9a6bc2 doc: reorganize prog guide 00:01:05.052 cb9187bc5c version: 24.11-rc0 00:01:05.052 b3485f4293 version: 24.07.0 00:01:05.052 fa58aec335 doc: add tested platforms with NVIDIA NICs 00:01:05.072 [Pipeline] writeFile 00:01:05.089 [Pipeline] sh 00:01:05.378 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:05.390 [Pipeline] sh 00:01:05.670 + cat autorun-spdk.conf 00:01:05.670 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:05.670 SPDK_RUN_ASAN=1 00:01:05.670 SPDK_RUN_UBSAN=1 00:01:05.670 SPDK_TEST_RAID=1 00:01:05.670 SPDK_TEST_NATIVE_DPDK=main 00:01:05.670 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:05.670 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:05.677 RUN_NIGHTLY=1 00:01:05.679 [Pipeline] } 00:01:05.693 [Pipeline] // stage 00:01:05.707 [Pipeline] stage 00:01:05.709 [Pipeline] { (Run VM) 00:01:05.722 [Pipeline] sh 00:01:06.008 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:06.008 + echo 'Start stage prepare_nvme.sh' 00:01:06.008 Start stage prepare_nvme.sh 00:01:06.008 + [[ -n 6 ]] 00:01:06.008 + disk_prefix=ex6 00:01:06.008 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:06.008 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:06.008 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:06.008 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:06.008 ++ SPDK_RUN_ASAN=1 00:01:06.008 ++ SPDK_RUN_UBSAN=1 00:01:06.008 ++ SPDK_TEST_RAID=1 00:01:06.008 ++ SPDK_TEST_NATIVE_DPDK=main 00:01:06.008 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:06.008 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:06.008 ++ RUN_NIGHTLY=1 00:01:06.008 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:06.008 + nvme_files=() 00:01:06.008 + declare -A nvme_files 00:01:06.008 + backend_dir=/var/lib/libvirt/images/backends 00:01:06.008 + nvme_files['nvme.img']=5G 00:01:06.008 + nvme_files['nvme-cmb.img']=5G 00:01:06.008 + nvme_files['nvme-multi0.img']=4G 00:01:06.008 + nvme_files['nvme-multi1.img']=4G 00:01:06.008 + nvme_files['nvme-multi2.img']=4G 00:01:06.008 + nvme_files['nvme-openstack.img']=8G 00:01:06.008 + nvme_files['nvme-zns.img']=5G 00:01:06.008 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:06.008 + (( SPDK_TEST_FTL == 1 )) 00:01:06.008 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:06.008 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi2.img -s 4G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-cmb.img -s 5G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-openstack.img -s 8G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-zns.img -s 5G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi1.img -s 4G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi0.img -s 4G 00:01:06.008 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:06.008 + for nvme in "${!nvme_files[@]}" 00:01:06.008 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme.img -s 5G 00:01:06.267 Formatting '/var/lib/libvirt/images/backends/ex6-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:06.267 ++ sudo grep -rl ex6-nvme.img /etc/libvirt/qemu 00:01:06.267 + echo 'End stage prepare_nvme.sh' 00:01:06.267 End stage prepare_nvme.sh 00:01:06.278 [Pipeline] sh 00:01:06.561 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:06.561 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex6-nvme.img -b /var/lib/libvirt/images/backends/ex6-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img -H -a -v -f fedora39 00:01:06.561 00:01:06.561 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:06.561 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:06.561 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:06.561 HELP=0 00:01:06.561 DRY_RUN=0 00:01:06.561 NVME_FILE=/var/lib/libvirt/images/backends/ex6-nvme.img,/var/lib/libvirt/images/backends/ex6-nvme-multi0.img, 00:01:06.561 NVME_DISKS_TYPE=nvme,nvme, 00:01:06.561 NVME_AUTO_CREATE=0 00:01:06.561 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img, 00:01:06.561 NVME_CMB=,, 00:01:06.561 NVME_PMR=,, 00:01:06.561 NVME_ZNS=,, 00:01:06.561 NVME_MS=,, 00:01:06.561 NVME_FDP=,, 00:01:06.561 SPDK_VAGRANT_DISTRO=fedora39 00:01:06.561 SPDK_VAGRANT_VMCPU=10 00:01:06.561 SPDK_VAGRANT_VMRAM=12288 00:01:06.561 SPDK_VAGRANT_PROVIDER=libvirt 00:01:06.561 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:06.561 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:06.561 SPDK_OPENSTACK_NETWORK=0 00:01:06.561 VAGRANT_PACKAGE_BOX=0 00:01:06.561 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:06.561 FORCE_DISTRO=true 00:01:06.561 VAGRANT_BOX_VERSION= 00:01:06.561 EXTRA_VAGRANTFILES= 00:01:06.561 NIC_MODEL=virtio 00:01:06.561 00:01:06.561 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:06.561 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:08.470 Bringing machine 'default' up with 'libvirt' provider... 00:01:08.730 ==> default: Creating image (snapshot of base box volume). 00:01:08.991 ==> default: Creating domain with the following settings... 00:01:08.991 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1727704813_3f161e2522deafeafc02 00:01:08.991 ==> default: -- Domain type: kvm 00:01:08.991 ==> default: -- Cpus: 10 00:01:08.991 ==> default: -- Feature: acpi 00:01:08.991 ==> default: -- Feature: apic 00:01:08.991 ==> default: -- Feature: pae 00:01:08.991 ==> default: -- Memory: 12288M 00:01:08.991 ==> default: -- Memory Backing: hugepages: 00:01:08.991 ==> default: -- Management MAC: 00:01:08.991 ==> default: -- Loader: 00:01:08.991 ==> default: -- Nvram: 00:01:08.991 ==> default: -- Base box: spdk/fedora39 00:01:08.991 ==> default: -- Storage pool: default 00:01:08.991 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1727704813_3f161e2522deafeafc02.img (20G) 00:01:08.991 ==> default: -- Volume Cache: default 00:01:08.991 ==> default: -- Kernel: 00:01:08.991 ==> default: -- Initrd: 00:01:08.991 ==> default: -- Graphics Type: vnc 00:01:08.991 ==> default: -- Graphics Port: -1 00:01:08.991 ==> default: -- Graphics IP: 127.0.0.1 00:01:08.991 ==> default: -- Graphics Password: Not defined 00:01:08.991 ==> default: -- Video Type: cirrus 00:01:08.991 ==> default: -- Video VRAM: 9216 00:01:08.991 ==> default: -- Sound Type: 00:01:08.991 ==> default: -- Keymap: en-us 00:01:08.991 ==> default: -- TPM Path: 00:01:08.991 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:08.991 ==> default: -- Command line args: 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:08.991 ==> default: -> value=-drive, 00:01:08.991 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme.img,if=none,id=nvme-0-drive0, 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:08.991 ==> default: -> value=-drive, 00:01:08.991 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:08.991 ==> default: -> value=-drive, 00:01:08.991 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:08.991 ==> default: -> value=-drive, 00:01:08.991 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:08.991 ==> default: -> value=-device, 00:01:08.991 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:08.991 ==> default: Creating shared folders metadata... 00:01:08.991 ==> default: Starting domain. 00:01:10.901 ==> default: Waiting for domain to get an IP address... 00:01:29.007 ==> default: Waiting for SSH to become available... 00:01:29.007 ==> default: Configuring and enabling network interfaces... 00:01:34.290 default: SSH address: 192.168.121.129:22 00:01:34.290 default: SSH username: vagrant 00:01:34.290 default: SSH auth method: private key 00:01:36.211 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:01:44.343 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:01:50.919 ==> default: Mounting SSHFS shared folder... 00:01:52.826 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:01:52.826 ==> default: Checking Mount.. 00:01:54.206 ==> default: Folder Successfully Mounted! 00:01:54.206 ==> default: Running provisioner: file... 00:01:55.586 default: ~/.gitconfig => .gitconfig 00:01:56.156 00:01:56.156 SUCCESS! 00:01:56.156 00:01:56.156 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:01:56.156 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:01:56.156 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:01:56.156 00:01:56.166 [Pipeline] } 00:01:56.180 [Pipeline] // stage 00:01:56.188 [Pipeline] dir 00:01:56.189 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:01:56.190 [Pipeline] { 00:01:56.202 [Pipeline] catchError 00:01:56.204 [Pipeline] { 00:01:56.215 [Pipeline] sh 00:01:56.498 + vagrant ssh-config --host vagrant 00:01:56.498 + sed -ne /^Host/,$p 00:01:56.498 + tee ssh_conf 00:01:59.036 Host vagrant 00:01:59.036 HostName 192.168.121.129 00:01:59.036 User vagrant 00:01:59.036 Port 22 00:01:59.036 UserKnownHostsFile /dev/null 00:01:59.036 StrictHostKeyChecking no 00:01:59.036 PasswordAuthentication no 00:01:59.036 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:01:59.036 IdentitiesOnly yes 00:01:59.036 LogLevel FATAL 00:01:59.036 ForwardAgent yes 00:01:59.036 ForwardX11 yes 00:01:59.036 00:01:59.050 [Pipeline] withEnv 00:01:59.052 [Pipeline] { 00:01:59.064 [Pipeline] sh 00:01:59.359 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:01:59.359 source /etc/os-release 00:01:59.359 [[ -e /image.version ]] && img=$(< /image.version) 00:01:59.359 # Minimal, systemd-like check. 00:01:59.359 if [[ -e /.dockerenv ]]; then 00:01:59.359 # Clear garbage from the node's name: 00:01:59.359 # agt-er_autotest_547-896 -> autotest_547-896 00:01:59.359 # $HOSTNAME is the actual container id 00:01:59.359 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:01:59.359 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:01:59.359 # We can assume this is a mount from a host where container is running, 00:01:59.359 # so fetch its hostname to easily identify the target swarm worker. 00:01:59.359 container="$(< /etc/hostname) ($agent)" 00:01:59.359 else 00:01:59.359 # Fallback 00:01:59.359 container=$agent 00:01:59.359 fi 00:01:59.359 fi 00:01:59.359 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:01:59.359 00:01:59.680 [Pipeline] } 00:01:59.694 [Pipeline] // withEnv 00:01:59.702 [Pipeline] setCustomBuildProperty 00:01:59.716 [Pipeline] stage 00:01:59.718 [Pipeline] { (Tests) 00:01:59.735 [Pipeline] sh 00:02:00.018 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:00.292 [Pipeline] sh 00:02:00.578 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:00.854 [Pipeline] timeout 00:02:00.854 Timeout set to expire in 1 hr 30 min 00:02:00.856 [Pipeline] { 00:02:00.869 [Pipeline] sh 00:02:01.153 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:01.722 HEAD is now at 09cc66129 test/unit: add mixed busy/idle mock poller function in reactor_ut 00:02:01.735 [Pipeline] sh 00:02:02.022 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:02.297 [Pipeline] sh 00:02:02.582 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:02.861 [Pipeline] sh 00:02:03.146 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:03.406 ++ readlink -f spdk_repo 00:02:03.406 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:03.406 + [[ -n /home/vagrant/spdk_repo ]] 00:02:03.406 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:03.406 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:03.406 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:03.406 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:03.406 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:03.406 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:03.406 + cd /home/vagrant/spdk_repo 00:02:03.406 + source /etc/os-release 00:02:03.406 ++ NAME='Fedora Linux' 00:02:03.406 ++ VERSION='39 (Cloud Edition)' 00:02:03.406 ++ ID=fedora 00:02:03.406 ++ VERSION_ID=39 00:02:03.406 ++ VERSION_CODENAME= 00:02:03.406 ++ PLATFORM_ID=platform:f39 00:02:03.406 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:03.406 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:03.406 ++ LOGO=fedora-logo-icon 00:02:03.406 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:03.406 ++ HOME_URL=https://fedoraproject.org/ 00:02:03.406 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:03.406 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:03.406 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:03.406 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:03.406 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:03.406 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:03.406 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:03.406 ++ SUPPORT_END=2024-11-12 00:02:03.406 ++ VARIANT='Cloud Edition' 00:02:03.406 ++ VARIANT_ID=cloud 00:02:03.406 + uname -a 00:02:03.406 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:03.406 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:03.978 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:03.978 Hugepages 00:02:03.978 node hugesize free / total 00:02:03.978 node0 1048576kB 0 / 0 00:02:03.978 node0 2048kB 0 / 0 00:02:03.978 00:02:03.978 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:03.978 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:03.978 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:02:04.239 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:02:04.239 + rm -f /tmp/spdk-ld-path 00:02:04.239 + source autorun-spdk.conf 00:02:04.239 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:04.239 ++ SPDK_RUN_ASAN=1 00:02:04.239 ++ SPDK_RUN_UBSAN=1 00:02:04.239 ++ SPDK_TEST_RAID=1 00:02:04.239 ++ SPDK_TEST_NATIVE_DPDK=main 00:02:04.239 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:04.239 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:04.239 ++ RUN_NIGHTLY=1 00:02:04.239 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:04.239 + [[ -n '' ]] 00:02:04.239 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:04.239 + for M in /var/spdk/build-*-manifest.txt 00:02:04.239 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:04.239 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:04.239 + for M in /var/spdk/build-*-manifest.txt 00:02:04.239 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:04.239 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:04.239 + for M in /var/spdk/build-*-manifest.txt 00:02:04.239 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:04.239 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:04.239 ++ uname 00:02:04.239 + [[ Linux == \L\i\n\u\x ]] 00:02:04.239 + sudo dmesg -T 00:02:04.239 + sudo dmesg --clear 00:02:04.239 + dmesg_pid=6169 00:02:04.239 + [[ Fedora Linux == FreeBSD ]] 00:02:04.239 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:04.239 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:04.239 + sudo dmesg -Tw 00:02:04.239 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:04.239 + [[ -x /usr/src/fio-static/fio ]] 00:02:04.239 + export FIO_BIN=/usr/src/fio-static/fio 00:02:04.239 + FIO_BIN=/usr/src/fio-static/fio 00:02:04.239 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:04.239 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:04.239 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:04.239 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:04.239 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:04.239 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:04.239 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:04.239 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:04.239 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:04.239 Test configuration: 00:02:04.239 SPDK_RUN_FUNCTIONAL_TEST=1 00:02:04.239 SPDK_RUN_ASAN=1 00:02:04.239 SPDK_RUN_UBSAN=1 00:02:04.239 SPDK_TEST_RAID=1 00:02:04.239 SPDK_TEST_NATIVE_DPDK=main 00:02:04.239 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:04.239 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:04.506 RUN_NIGHTLY=1 14:01:08 -- common/autotest_common.sh@1680 -- $ [[ n == y ]] 00:02:04.506 14:01:08 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:04.506 14:01:08 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:04.506 14:01:08 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:04.506 14:01:08 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:04.506 14:01:08 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:04.506 14:01:08 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.506 14:01:08 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.506 14:01:08 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.506 14:01:08 -- paths/export.sh@5 -- $ export PATH 00:02:04.506 14:01:08 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.506 14:01:08 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:04.506 14:01:08 -- common/autobuild_common.sh@479 -- $ date +%s 00:02:04.506 14:01:08 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727704868.XXXXXX 00:02:04.506 14:01:08 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727704868.k2VwBB 00:02:04.506 14:01:08 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:02:04.506 14:01:08 -- common/autobuild_common.sh@485 -- $ '[' -n main ']' 00:02:04.506 14:01:08 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:04.506 14:01:08 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:04.506 14:01:08 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:04.506 14:01:08 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:04.506 14:01:08 -- common/autobuild_common.sh@495 -- $ get_config_params 00:02:04.506 14:01:08 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:02:04.506 14:01:08 -- common/autotest_common.sh@10 -- $ set +x 00:02:04.506 14:01:09 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:04.506 14:01:09 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:02:04.506 14:01:09 -- pm/common@17 -- $ local monitor 00:02:04.506 14:01:09 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.506 14:01:09 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.506 14:01:09 -- pm/common@25 -- $ sleep 1 00:02:04.506 14:01:09 -- pm/common@21 -- $ date +%s 00:02:04.506 14:01:09 -- pm/common@21 -- $ date +%s 00:02:04.506 14:01:09 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727704869 00:02:04.506 14:01:09 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727704869 00:02:04.506 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727704869_collect-vmstat.pm.log 00:02:04.506 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727704869_collect-cpu-load.pm.log 00:02:05.445 14:01:10 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:02:05.445 14:01:10 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:05.445 14:01:10 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:05.445 14:01:10 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:05.445 14:01:10 -- spdk/autobuild.sh@16 -- $ date -u 00:02:05.445 Mon Sep 30 02:01:10 PM UTC 2024 00:02:05.445 14:01:10 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:05.445 v25.01-pre-17-g09cc66129 00:02:05.445 14:01:10 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:05.445 14:01:10 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:05.445 14:01:10 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:05.445 14:01:10 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:05.445 14:01:10 -- common/autotest_common.sh@10 -- $ set +x 00:02:05.445 ************************************ 00:02:05.445 START TEST asan 00:02:05.445 ************************************ 00:02:05.445 using asan 00:02:05.445 14:01:10 asan -- common/autotest_common.sh@1125 -- $ echo 'using asan' 00:02:05.445 00:02:05.445 real 0m0.000s 00:02:05.445 user 0m0.000s 00:02:05.445 sys 0m0.000s 00:02:05.445 14:01:10 asan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:05.445 14:01:10 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:05.445 ************************************ 00:02:05.445 END TEST asan 00:02:05.445 ************************************ 00:02:05.705 14:01:10 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:05.705 14:01:10 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:05.705 14:01:10 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:05.705 14:01:10 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:05.705 14:01:10 -- common/autotest_common.sh@10 -- $ set +x 00:02:05.705 ************************************ 00:02:05.705 START TEST ubsan 00:02:05.705 ************************************ 00:02:05.705 using ubsan 00:02:05.705 14:01:10 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:02:05.705 00:02:05.706 real 0m0.000s 00:02:05.706 user 0m0.000s 00:02:05.706 sys 0m0.000s 00:02:05.706 14:01:10 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:05.706 14:01:10 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:05.706 ************************************ 00:02:05.706 END TEST ubsan 00:02:05.706 ************************************ 00:02:05.706 14:01:10 -- spdk/autobuild.sh@27 -- $ '[' -n main ']' 00:02:05.706 14:01:10 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:05.706 14:01:10 -- common/autobuild_common.sh@442 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:05.706 14:01:10 -- common/autotest_common.sh@1101 -- $ '[' 2 -le 1 ']' 00:02:05.706 14:01:10 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:05.706 14:01:10 -- common/autotest_common.sh@10 -- $ set +x 00:02:05.706 ************************************ 00:02:05.706 START TEST build_native_dpdk 00:02:05.706 ************************************ 00:02:05.706 14:01:10 build_native_dpdk -- common/autotest_common.sh@1125 -- $ _build_native_dpdk 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:05.706 bf0ff8df59 maintainers: fix prog guide paths 00:02:05.706 41dd9a6bc2 doc: reorganize prog guide 00:02:05.706 cb9187bc5c version: 24.11-rc0 00:02:05.706 b3485f4293 version: 24.07.0 00:02:05.706 fa58aec335 doc: add tested platforms with NVIDIA NICs 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=24.11.0-rc0 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@100 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base") 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@102 -- $ local mlx5_libs_added=n 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@139 -- $ [[ 0 -eq 1 ]] 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@167 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@168 -- $ uname -s 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@168 -- $ '[' Linux = Linux ']' 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@169 -- $ lt 24.11.0-rc0 21.11.0 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 24.11.0-rc0 '<' 21.11.0 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=4 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@173 -- $ patch -p1 00:02:05.706 patching file config/rte_config.h 00:02:05.706 Hunk #1 succeeded at 70 (offset 11 lines). 00:02:05.706 14:01:10 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 24.11.0-rc0 24.07.0 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 24.11.0-rc0 '<' 24.07.0 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=4 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:05.706 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v++ )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 11 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 07 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=07 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 07 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 7 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=7 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:05.707 14:01:10 build_native_dpdk -- common/autobuild_common.sh@179 -- $ ge 24.11.0-rc0 24.07.0 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 24.11.0-rc0 '>=' 24.07.0 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=4 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v++ )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 11 =~ ^[0-9]+$ ]] 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=11 00:02:05.707 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 07 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@353 -- $ local d=07 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 07 =~ ^[0-9]+$ ]] 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@355 -- $ echo 7 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=7 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:05.967 14:01:10 build_native_dpdk -- scripts/common.sh@367 -- $ return 0 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:02:05.967 patching file drivers/bus/pci/linux/pci_uio.c 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@183 -- $ dpdk_kmods=false 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@184 -- $ uname -s 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@184 -- $ '[' Linux = FreeBSD ']' 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@188 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base 00:02:05.967 14:01:10 build_native_dpdk -- common/autobuild_common.sh@188 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:12.545 The Meson build system 00:02:12.545 Version: 1.5.0 00:02:12.545 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:12.545 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:12.545 Build type: native build 00:02:12.545 Program cat found: YES (/usr/bin/cat) 00:02:12.545 Project name: DPDK 00:02:12.545 Project version: 24.11.0-rc0 00:02:12.545 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:12.545 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:12.545 Host machine cpu family: x86_64 00:02:12.545 Host machine cpu: x86_64 00:02:12.545 Message: ## Building in Developer Mode ## 00:02:12.545 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:12.545 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:12.545 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:12.545 Program python3 (elftools) found: YES (/usr/bin/python3) modules: elftools 00:02:12.545 Program cat found: YES (/usr/bin/cat) 00:02:12.545 config/meson.build:120: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:12.545 Compiler for C supports arguments -march=native: YES 00:02:12.545 Checking for size of "void *" : 8 00:02:12.545 Checking for size of "void *" : 8 (cached) 00:02:12.545 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:02:12.545 Library m found: YES 00:02:12.545 Library numa found: YES 00:02:12.545 Has header "numaif.h" : YES 00:02:12.545 Library fdt found: NO 00:02:12.545 Library execinfo found: NO 00:02:12.545 Has header "execinfo.h" : YES 00:02:12.545 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:12.545 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:12.545 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:12.545 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:12.545 Run-time dependency openssl found: YES 3.1.1 00:02:12.545 Run-time dependency libpcap found: YES 1.10.4 00:02:12.545 Has header "pcap.h" with dependency libpcap: YES 00:02:12.545 Compiler for C supports arguments -Wcast-qual: YES 00:02:12.545 Compiler for C supports arguments -Wdeprecated: YES 00:02:12.545 Compiler for C supports arguments -Wformat: YES 00:02:12.545 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:12.545 Compiler for C supports arguments -Wformat-security: NO 00:02:12.545 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:12.546 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:12.546 Compiler for C supports arguments -Wnested-externs: YES 00:02:12.546 Compiler for C supports arguments -Wold-style-definition: YES 00:02:12.546 Compiler for C supports arguments -Wpointer-arith: YES 00:02:12.546 Compiler for C supports arguments -Wsign-compare: YES 00:02:12.546 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:12.546 Compiler for C supports arguments -Wundef: YES 00:02:12.546 Compiler for C supports arguments -Wwrite-strings: YES 00:02:12.546 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:12.546 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:12.546 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:12.546 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:12.546 Program objdump found: YES (/usr/bin/objdump) 00:02:12.546 Compiler for C supports arguments -mavx512f: YES 00:02:12.546 Checking if "AVX512 checking" compiles: YES 00:02:12.546 Fetching value of define "__SSE4_2__" : 1 00:02:12.546 Fetching value of define "__AES__" : 1 00:02:12.546 Fetching value of define "__AVX__" : 1 00:02:12.546 Fetching value of define "__AVX2__" : 1 00:02:12.546 Fetching value of define "__AVX512BW__" : 1 00:02:12.546 Fetching value of define "__AVX512CD__" : 1 00:02:12.546 Fetching value of define "__AVX512DQ__" : 1 00:02:12.546 Fetching value of define "__AVX512F__" : 1 00:02:12.546 Fetching value of define "__AVX512VL__" : 1 00:02:12.546 Fetching value of define "__PCLMUL__" : 1 00:02:12.546 Fetching value of define "__RDRND__" : 1 00:02:12.546 Fetching value of define "__RDSEED__" : 1 00:02:12.546 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:12.546 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:12.546 Message: lib/log: Defining dependency "log" 00:02:12.546 Message: lib/kvargs: Defining dependency "kvargs" 00:02:12.546 Message: lib/argparse: Defining dependency "argparse" 00:02:12.546 Message: lib/telemetry: Defining dependency "telemetry" 00:02:12.546 Checking for function "getentropy" : NO 00:02:12.546 Message: lib/eal: Defining dependency "eal" 00:02:12.546 Message: lib/ptr_compress: Defining dependency "ptr_compress" 00:02:12.546 Message: lib/ring: Defining dependency "ring" 00:02:12.546 Message: lib/rcu: Defining dependency "rcu" 00:02:12.546 Message: lib/mempool: Defining dependency "mempool" 00:02:12.546 Message: lib/mbuf: Defining dependency "mbuf" 00:02:12.546 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:12.546 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:12.546 Compiler for C supports arguments -mpclmul: YES 00:02:12.546 Compiler for C supports arguments -maes: YES 00:02:12.546 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:12.546 Compiler for C supports arguments -mavx512bw: YES 00:02:12.546 Compiler for C supports arguments -mavx512dq: YES 00:02:12.546 Compiler for C supports arguments -mavx512vl: YES 00:02:12.546 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:12.546 Compiler for C supports arguments -mavx2: YES 00:02:12.546 Compiler for C supports arguments -mavx: YES 00:02:12.546 Message: lib/net: Defining dependency "net" 00:02:12.546 Message: lib/meter: Defining dependency "meter" 00:02:12.546 Message: lib/ethdev: Defining dependency "ethdev" 00:02:12.546 Message: lib/pci: Defining dependency "pci" 00:02:12.546 Message: lib/cmdline: Defining dependency "cmdline" 00:02:12.546 Message: lib/metrics: Defining dependency "metrics" 00:02:12.546 Message: lib/hash: Defining dependency "hash" 00:02:12.546 Message: lib/timer: Defining dependency "timer" 00:02:12.546 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:12.546 Message: lib/acl: Defining dependency "acl" 00:02:12.546 Message: lib/bbdev: Defining dependency "bbdev" 00:02:12.546 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:12.546 Run-time dependency libelf found: YES 0.191 00:02:12.546 Message: lib/bpf: Defining dependency "bpf" 00:02:12.546 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:12.546 Message: lib/compressdev: Defining dependency "compressdev" 00:02:12.546 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:12.546 Message: lib/distributor: Defining dependency "distributor" 00:02:12.546 Message: lib/dmadev: Defining dependency "dmadev" 00:02:12.546 Message: lib/efd: Defining dependency "efd" 00:02:12.546 Message: lib/eventdev: Defining dependency "eventdev" 00:02:12.546 Message: lib/dispatcher: Defining dependency "dispatcher" 00:02:12.546 Message: lib/gpudev: Defining dependency "gpudev" 00:02:12.546 Message: lib/gro: Defining dependency "gro" 00:02:12.546 Message: lib/gso: Defining dependency "gso" 00:02:12.546 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:12.546 Message: lib/jobstats: Defining dependency "jobstats" 00:02:12.546 Message: lib/latencystats: Defining dependency "latencystats" 00:02:12.546 Message: lib/lpm: Defining dependency "lpm" 00:02:12.546 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:12.546 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:12.546 Message: lib/member: Defining dependency "member" 00:02:12.546 Message: lib/pcapng: Defining dependency "pcapng" 00:02:12.546 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:12.546 Message: lib/power: Defining dependency "power" 00:02:12.546 Message: lib/rawdev: Defining dependency "rawdev" 00:02:12.546 Message: lib/regexdev: Defining dependency "regexdev" 00:02:12.546 Message: lib/mldev: Defining dependency "mldev" 00:02:12.546 Message: lib/rib: Defining dependency "rib" 00:02:12.546 Message: lib/reorder: Defining dependency "reorder" 00:02:12.546 Message: lib/sched: Defining dependency "sched" 00:02:12.546 Message: lib/security: Defining dependency "security" 00:02:12.546 Message: lib/stack: Defining dependency "stack" 00:02:12.546 Has header "linux/userfaultfd.h" : YES 00:02:12.546 Has header "linux/vduse.h" : YES 00:02:12.546 Message: lib/vhost: Defining dependency "vhost" 00:02:12.546 Message: lib/ipsec: Defining dependency "ipsec" 00:02:12.546 Message: lib/pdcp: Defining dependency "pdcp" 00:02:12.546 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:12.546 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:12.546 Message: lib/fib: Defining dependency "fib" 00:02:12.546 Message: lib/port: Defining dependency "port" 00:02:12.546 Message: lib/pdump: Defining dependency "pdump" 00:02:12.546 Message: lib/table: Defining dependency "table" 00:02:12.546 Message: lib/pipeline: Defining dependency "pipeline" 00:02:12.546 Message: lib/graph: Defining dependency "graph" 00:02:12.546 Message: lib/node: Defining dependency "node" 00:02:12.546 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:12.546 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:12.546 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:13.116 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:13.116 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:13.116 Compiler for C supports arguments -Wno-unused-value: YES 00:02:13.116 Compiler for C supports arguments -Wno-format: YES 00:02:13.116 Compiler for C supports arguments -Wno-format-security: YES 00:02:13.116 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:13.116 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:13.116 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:13.116 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:13.116 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:13.116 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:13.116 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:13.116 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:13.116 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:13.116 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:13.116 Has header "sys/epoll.h" : YES 00:02:13.116 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:13.116 Configuring doxy-api-html.conf using configuration 00:02:13.116 Configuring doxy-api-man.conf using configuration 00:02:13.116 Program mandb found: YES (/usr/bin/mandb) 00:02:13.116 Program sphinx-build found: NO 00:02:13.116 Configuring rte_build_config.h using configuration 00:02:13.116 Message: 00:02:13.116 ================= 00:02:13.116 Applications Enabled 00:02:13.116 ================= 00:02:13.116 00:02:13.116 apps: 00:02:13.116 dumpcap, graph, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, 00:02:13.116 test-crypto-perf, test-dma-perf, test-eventdev, test-fib, test-flow-perf, test-gpudev, test-mldev, test-pipeline, 00:02:13.116 test-pmd, test-regex, test-sad, test-security-perf, 00:02:13.116 00:02:13.116 Message: 00:02:13.116 ================= 00:02:13.116 Libraries Enabled 00:02:13.116 ================= 00:02:13.116 00:02:13.116 libs: 00:02:13.116 log, kvargs, argparse, telemetry, eal, ptr_compress, ring, rcu, 00:02:13.116 mempool, mbuf, net, meter, ethdev, pci, cmdline, metrics, 00:02:13.116 hash, timer, acl, bbdev, bitratestats, bpf, cfgfile, compressdev, 00:02:13.116 cryptodev, distributor, dmadev, efd, eventdev, dispatcher, gpudev, gro, 00:02:13.116 gso, ip_frag, jobstats, latencystats, lpm, member, pcapng, power, 00:02:13.116 rawdev, regexdev, mldev, rib, reorder, sched, security, stack, 00:02:13.116 vhost, ipsec, pdcp, fib, port, pdump, table, pipeline, 00:02:13.116 graph, node, 00:02:13.116 00:02:13.116 Message: 00:02:13.116 =============== 00:02:13.116 Drivers Enabled 00:02:13.116 =============== 00:02:13.116 00:02:13.116 common: 00:02:13.116 00:02:13.116 bus: 00:02:13.116 pci, vdev, 00:02:13.116 mempool: 00:02:13.116 ring, 00:02:13.116 dma: 00:02:13.116 00:02:13.116 net: 00:02:13.116 i40e, 00:02:13.116 raw: 00:02:13.116 00:02:13.116 crypto: 00:02:13.116 00:02:13.116 compress: 00:02:13.116 00:02:13.116 regex: 00:02:13.116 00:02:13.116 ml: 00:02:13.116 00:02:13.116 vdpa: 00:02:13.116 00:02:13.116 event: 00:02:13.116 00:02:13.116 baseband: 00:02:13.116 00:02:13.116 gpu: 00:02:13.116 00:02:13.116 00:02:13.116 Message: 00:02:13.116 ================= 00:02:13.116 Content Skipped 00:02:13.116 ================= 00:02:13.116 00:02:13.116 apps: 00:02:13.116 00:02:13.116 libs: 00:02:13.116 00:02:13.116 drivers: 00:02:13.116 common/cpt: not in enabled drivers build config 00:02:13.116 common/dpaax: not in enabled drivers build config 00:02:13.116 common/iavf: not in enabled drivers build config 00:02:13.116 common/idpf: not in enabled drivers build config 00:02:13.116 common/ionic: not in enabled drivers build config 00:02:13.116 common/mvep: not in enabled drivers build config 00:02:13.116 common/octeontx: not in enabled drivers build config 00:02:13.116 bus/auxiliary: not in enabled drivers build config 00:02:13.116 bus/cdx: not in enabled drivers build config 00:02:13.116 bus/dpaa: not in enabled drivers build config 00:02:13.116 bus/fslmc: not in enabled drivers build config 00:02:13.116 bus/ifpga: not in enabled drivers build config 00:02:13.116 bus/platform: not in enabled drivers build config 00:02:13.116 bus/uacce: not in enabled drivers build config 00:02:13.116 bus/vmbus: not in enabled drivers build config 00:02:13.116 common/cnxk: not in enabled drivers build config 00:02:13.116 common/mlx5: not in enabled drivers build config 00:02:13.116 common/nfp: not in enabled drivers build config 00:02:13.116 common/nitrox: not in enabled drivers build config 00:02:13.116 common/qat: not in enabled drivers build config 00:02:13.116 common/sfc_efx: not in enabled drivers build config 00:02:13.116 mempool/bucket: not in enabled drivers build config 00:02:13.116 mempool/cnxk: not in enabled drivers build config 00:02:13.116 mempool/dpaa: not in enabled drivers build config 00:02:13.116 mempool/dpaa2: not in enabled drivers build config 00:02:13.116 mempool/octeontx: not in enabled drivers build config 00:02:13.116 mempool/stack: not in enabled drivers build config 00:02:13.116 dma/cnxk: not in enabled drivers build config 00:02:13.116 dma/dpaa: not in enabled drivers build config 00:02:13.116 dma/dpaa2: not in enabled drivers build config 00:02:13.116 dma/hisilicon: not in enabled drivers build config 00:02:13.116 dma/idxd: not in enabled drivers build config 00:02:13.116 dma/ioat: not in enabled drivers build config 00:02:13.116 dma/odm: not in enabled drivers build config 00:02:13.116 dma/skeleton: not in enabled drivers build config 00:02:13.116 net/af_packet: not in enabled drivers build config 00:02:13.116 net/af_xdp: not in enabled drivers build config 00:02:13.116 net/ark: not in enabled drivers build config 00:02:13.116 net/atlantic: not in enabled drivers build config 00:02:13.116 net/avp: not in enabled drivers build config 00:02:13.116 net/axgbe: not in enabled drivers build config 00:02:13.116 net/bnx2x: not in enabled drivers build config 00:02:13.116 net/bnxt: not in enabled drivers build config 00:02:13.116 net/bonding: not in enabled drivers build config 00:02:13.116 net/cnxk: not in enabled drivers build config 00:02:13.116 net/cpfl: not in enabled drivers build config 00:02:13.116 net/cxgbe: not in enabled drivers build config 00:02:13.116 net/dpaa: not in enabled drivers build config 00:02:13.116 net/dpaa2: not in enabled drivers build config 00:02:13.116 net/e1000: not in enabled drivers build config 00:02:13.116 net/ena: not in enabled drivers build config 00:02:13.116 net/enetc: not in enabled drivers build config 00:02:13.116 net/enetfec: not in enabled drivers build config 00:02:13.116 net/enic: not in enabled drivers build config 00:02:13.116 net/failsafe: not in enabled drivers build config 00:02:13.116 net/fm10k: not in enabled drivers build config 00:02:13.116 net/gve: not in enabled drivers build config 00:02:13.116 net/hinic: not in enabled drivers build config 00:02:13.116 net/hns3: not in enabled drivers build config 00:02:13.116 net/iavf: not in enabled drivers build config 00:02:13.116 net/ice: not in enabled drivers build config 00:02:13.116 net/idpf: not in enabled drivers build config 00:02:13.116 net/igc: not in enabled drivers build config 00:02:13.116 net/ionic: not in enabled drivers build config 00:02:13.116 net/ipn3ke: not in enabled drivers build config 00:02:13.116 net/ixgbe: not in enabled drivers build config 00:02:13.116 net/mana: not in enabled drivers build config 00:02:13.116 net/memif: not in enabled drivers build config 00:02:13.116 net/mlx4: not in enabled drivers build config 00:02:13.116 net/mlx5: not in enabled drivers build config 00:02:13.116 net/mvneta: not in enabled drivers build config 00:02:13.116 net/mvpp2: not in enabled drivers build config 00:02:13.116 net/netvsc: not in enabled drivers build config 00:02:13.116 net/nfb: not in enabled drivers build config 00:02:13.116 net/nfp: not in enabled drivers build config 00:02:13.116 net/ngbe: not in enabled drivers build config 00:02:13.117 net/ntnic: not in enabled drivers build config 00:02:13.117 net/null: not in enabled drivers build config 00:02:13.117 net/octeontx: not in enabled drivers build config 00:02:13.117 net/octeon_ep: not in enabled drivers build config 00:02:13.117 net/pcap: not in enabled drivers build config 00:02:13.117 net/pfe: not in enabled drivers build config 00:02:13.117 net/qede: not in enabled drivers build config 00:02:13.117 net/ring: not in enabled drivers build config 00:02:13.117 net/sfc: not in enabled drivers build config 00:02:13.117 net/softnic: not in enabled drivers build config 00:02:13.117 net/tap: not in enabled drivers build config 00:02:13.117 net/thunderx: not in enabled drivers build config 00:02:13.117 net/txgbe: not in enabled drivers build config 00:02:13.117 net/vdev_netvsc: not in enabled drivers build config 00:02:13.117 net/vhost: not in enabled drivers build config 00:02:13.117 net/virtio: not in enabled drivers build config 00:02:13.117 net/vmxnet3: not in enabled drivers build config 00:02:13.117 raw/cnxk_bphy: not in enabled drivers build config 00:02:13.117 raw/cnxk_gpio: not in enabled drivers build config 00:02:13.117 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:13.117 raw/ifpga: not in enabled drivers build config 00:02:13.117 raw/ntb: not in enabled drivers build config 00:02:13.117 raw/skeleton: not in enabled drivers build config 00:02:13.117 crypto/armv8: not in enabled drivers build config 00:02:13.117 crypto/bcmfs: not in enabled drivers build config 00:02:13.117 crypto/caam_jr: not in enabled drivers build config 00:02:13.117 crypto/ccp: not in enabled drivers build config 00:02:13.117 crypto/cnxk: not in enabled drivers build config 00:02:13.117 crypto/dpaa_sec: not in enabled drivers build config 00:02:13.117 crypto/dpaa2_sec: not in enabled drivers build config 00:02:13.117 crypto/ionic: not in enabled drivers build config 00:02:13.117 crypto/ipsec_mb: not in enabled drivers build config 00:02:13.117 crypto/mlx5: not in enabled drivers build config 00:02:13.117 crypto/mvsam: not in enabled drivers build config 00:02:13.117 crypto/nitrox: not in enabled drivers build config 00:02:13.117 crypto/null: not in enabled drivers build config 00:02:13.117 crypto/octeontx: not in enabled drivers build config 00:02:13.117 crypto/openssl: not in enabled drivers build config 00:02:13.117 crypto/scheduler: not in enabled drivers build config 00:02:13.117 crypto/uadk: not in enabled drivers build config 00:02:13.117 crypto/virtio: not in enabled drivers build config 00:02:13.117 compress/isal: not in enabled drivers build config 00:02:13.117 compress/mlx5: not in enabled drivers build config 00:02:13.117 compress/nitrox: not in enabled drivers build config 00:02:13.117 compress/octeontx: not in enabled drivers build config 00:02:13.117 compress/uadk: not in enabled drivers build config 00:02:13.117 compress/zlib: not in enabled drivers build config 00:02:13.117 regex/mlx5: not in enabled drivers build config 00:02:13.117 regex/cn9k: not in enabled drivers build config 00:02:13.117 ml/cnxk: not in enabled drivers build config 00:02:13.117 vdpa/ifc: not in enabled drivers build config 00:02:13.117 vdpa/mlx5: not in enabled drivers build config 00:02:13.117 vdpa/nfp: not in enabled drivers build config 00:02:13.117 vdpa/sfc: not in enabled drivers build config 00:02:13.117 event/cnxk: not in enabled drivers build config 00:02:13.117 event/dlb2: not in enabled drivers build config 00:02:13.117 event/dpaa: not in enabled drivers build config 00:02:13.117 event/dpaa2: not in enabled drivers build config 00:02:13.117 event/dsw: not in enabled drivers build config 00:02:13.117 event/opdl: not in enabled drivers build config 00:02:13.117 event/skeleton: not in enabled drivers build config 00:02:13.117 event/sw: not in enabled drivers build config 00:02:13.117 event/octeontx: not in enabled drivers build config 00:02:13.117 baseband/acc: not in enabled drivers build config 00:02:13.117 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:13.117 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:13.117 baseband/la12xx: not in enabled drivers build config 00:02:13.117 baseband/null: not in enabled drivers build config 00:02:13.117 baseband/turbo_sw: not in enabled drivers build config 00:02:13.117 gpu/cuda: not in enabled drivers build config 00:02:13.117 00:02:13.117 00:02:13.117 Build targets in project: 221 00:02:13.117 00:02:13.117 DPDK 24.11.0-rc0 00:02:13.117 00:02:13.117 User defined options 00:02:13.117 libdir : lib 00:02:13.117 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:13.117 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:13.117 c_link_args : 00:02:13.117 enable_docs : false 00:02:13.117 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:13.117 enable_kmods : false 00:02:13.117 machine : native 00:02:13.117 tests : false 00:02:13.117 00:02:13.117 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:13.117 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:13.377 14:01:17 build_native_dpdk -- common/autobuild_common.sh@192 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:13.377 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:13.377 [1/720] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:02:13.377 [2/720] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:13.377 [3/720] Linking static target lib/librte_kvargs.a 00:02:13.377 [4/720] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:13.377 [5/720] Compiling C object lib/librte_log.a.p/log_log.c.o 00:02:13.636 [6/720] Linking static target lib/librte_log.a 00:02:13.636 [7/720] Compiling C object lib/librte_argparse.a.p/argparse_rte_argparse.c.o 00:02:13.636 [8/720] Linking static target lib/librte_argparse.a 00:02:13.636 [9/720] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:13.637 [10/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:13.637 [11/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:13.637 [12/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:13.637 [13/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:13.637 [14/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:13.637 [15/720] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:13.637 [16/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:13.896 [17/720] Generating lib/argparse.sym_chk with a custom command (wrapped by meson to capture output) 00:02:13.896 [18/720] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:02:13.896 [19/720] Linking target lib/librte_log.so.25.0 00:02:13.896 [20/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:13.896 [21/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:14.156 [22/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:14.156 [23/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:14.156 [24/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:14.156 [25/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:14.156 [26/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:14.156 [27/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:14.156 [28/720] Generating symbol file lib/librte_log.so.25.0.p/librte_log.so.25.0.symbols 00:02:14.156 [29/720] Linking target lib/librte_kvargs.so.25.0 00:02:14.156 [30/720] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:14.156 [31/720] Linking target lib/librte_argparse.so.25.0 00:02:14.156 [32/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:14.156 [33/720] Linking static target lib/librte_telemetry.a 00:02:14.416 [34/720] Generating symbol file lib/librte_kvargs.so.25.0.p/librte_kvargs.so.25.0.symbols 00:02:14.416 [35/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:14.416 [36/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:14.416 [37/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:14.416 [38/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:14.416 [39/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:14.675 [40/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:14.675 [41/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:14.675 [42/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:14.675 [43/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:14.675 [44/720] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:14.675 [45/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:14.675 [46/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:14.675 [47/720] Linking target lib/librte_telemetry.so.25.0 00:02:14.675 [48/720] Generating symbol file lib/librte_telemetry.so.25.0.p/librte_telemetry.so.25.0.symbols 00:02:14.675 [49/720] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:14.937 [50/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:14.937 [51/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:14.937 [52/720] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:14.937 [53/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:14.937 [54/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:14.937 [55/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:14.937 [56/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:15.198 [57/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:15.198 [58/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:15.198 [59/720] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:15.198 [60/720] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:15.198 [61/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:15.198 [62/720] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:15.198 [63/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:15.198 [64/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:15.458 [65/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:15.458 [66/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:15.458 [67/720] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:15.458 [68/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:15.458 [69/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:15.458 [70/720] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:15.458 [71/720] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:15.718 [72/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:15.718 [73/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:15.718 [74/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:15.718 [75/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:15.718 [76/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:15.718 [77/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:15.718 [78/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:15.718 [79/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:15.718 [80/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:15.978 [81/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:15.978 [82/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:15.978 [83/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:15.978 [84/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:15.978 [85/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_mmu.c.o 00:02:15.978 [86/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:16.238 [87/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:16.238 [88/720] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:16.238 [89/720] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:16.238 [90/720] Linking static target lib/librte_ring.a 00:02:16.498 [91/720] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:16.498 [92/720] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:16.498 [93/720] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:16.498 [94/720] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:16.498 [95/720] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:16.498 [96/720] Linking static target lib/librte_eal.a 00:02:16.498 [97/720] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:16.498 [98/720] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:16.759 [99/720] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:16.759 [100/720] Linking static target lib/librte_mempool.a 00:02:16.759 [101/720] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:16.759 [102/720] Linking static target lib/librte_rcu.a 00:02:16.759 [103/720] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:16.759 [104/720] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:16.759 [105/720] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:16.759 [106/720] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:16.759 [107/720] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:17.018 [108/720] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:17.018 [109/720] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:17.018 [110/720] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:17.018 [111/720] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:17.018 [112/720] Linking static target lib/librte_mbuf.a 00:02:17.278 [113/720] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:17.278 [114/720] Linking static target lib/librte_net.a 00:02:17.278 [115/720] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:17.278 [116/720] Linking static target lib/librte_meter.a 00:02:17.278 [117/720] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:17.278 [118/720] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:17.278 [119/720] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:17.278 [120/720] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:17.538 [121/720] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:17.538 [122/720] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:17.538 [123/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:17.538 [124/720] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:17.798 [125/720] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:17.798 [126/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:18.057 [127/720] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:18.057 [128/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:02:18.057 [129/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:18.057 [130/720] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:18.316 [131/720] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:18.316 [132/720] Linking static target lib/librte_pci.a 00:02:18.316 [133/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:18.316 [134/720] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:18.316 [135/720] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:18.316 [136/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:18.316 [137/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:18.316 [138/720] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:18.316 [139/720] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:02:18.316 [140/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:18.575 [141/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:18.575 [142/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:18.575 [143/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:18.575 [144/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:18.575 [145/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:18.575 [146/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:18.575 [147/720] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:18.575 [148/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:18.575 [149/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:18.834 [150/720] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:18.834 [151/720] Linking static target lib/librte_cmdline.a 00:02:18.834 [152/720] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:02:18.834 [153/720] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:18.834 [154/720] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:18.834 [155/720] Linking static target lib/librte_metrics.a 00:02:18.834 [156/720] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:18.834 [157/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:19.093 [158/720] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:02:19.093 [159/720] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:19.352 [160/720] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:19.352 [161/720] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:19.352 [162/720] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:19.352 [163/720] Linking static target lib/librte_timer.a 00:02:19.611 [164/720] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:19.611 [165/720] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:19.611 [166/720] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:19.611 [167/720] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:19.611 [168/720] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:20.180 [169/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:20.180 [170/720] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:20.180 [171/720] Linking static target lib/librte_bitratestats.a 00:02:20.180 [172/720] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:20.438 [173/720] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:20.438 [174/720] Linking static target lib/librte_bbdev.a 00:02:20.438 [175/720] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:20.438 [176/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:20.698 [177/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:20.698 [178/720] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:20.698 [179/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:20.698 [180/720] Linking static target lib/librte_hash.a 00:02:20.956 [181/720] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:20.957 [182/720] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:20.957 [183/720] Linking static target lib/librte_ethdev.a 00:02:20.957 [184/720] Compiling C object lib/acl/libavx2_tmp.a.p/acl_run_avx2.c.o 00:02:20.957 [185/720] Linking static target lib/acl/libavx2_tmp.a 00:02:21.216 [186/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:21.216 [187/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:21.216 [188/720] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:21.216 [189/720] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.216 [190/720] Linking target lib/librte_eal.so.25.0 00:02:21.475 [191/720] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.475 [192/720] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:21.475 [193/720] Linking static target lib/librte_cfgfile.a 00:02:21.475 [194/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:21.475 [195/720] Generating symbol file lib/librte_eal.so.25.0.p/librte_eal.so.25.0.symbols 00:02:21.475 [196/720] Linking target lib/librte_ring.so.25.0 00:02:21.475 [197/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:21.475 [198/720] Generating symbol file lib/librte_ring.so.25.0.p/librte_ring.so.25.0.symbols 00:02:21.735 [199/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:21.735 [200/720] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.735 [201/720] Linking target lib/librte_rcu.so.25.0 00:02:21.735 [202/720] Linking target lib/librte_meter.so.25.0 00:02:21.735 [203/720] Linking target lib/librte_mempool.so.25.0 00:02:21.735 [204/720] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:21.735 [205/720] Linking target lib/librte_pci.so.25.0 00:02:21.735 [206/720] Linking target lib/librte_timer.so.25.0 00:02:21.735 [207/720] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:21.735 [208/720] Linking target lib/librte_cfgfile.so.25.0 00:02:21.735 [209/720] Generating symbol file lib/librte_rcu.so.25.0.p/librte_rcu.so.25.0.symbols 00:02:21.735 [210/720] Generating symbol file lib/librte_mempool.so.25.0.p/librte_mempool.so.25.0.symbols 00:02:21.735 [211/720] Generating symbol file lib/librte_pci.so.25.0.p/librte_pci.so.25.0.symbols 00:02:21.735 [212/720] Generating symbol file lib/librte_meter.so.25.0.p/librte_meter.so.25.0.symbols 00:02:21.735 [213/720] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:21.735 [214/720] Generating symbol file lib/librte_timer.so.25.0.p/librte_timer.so.25.0.symbols 00:02:21.735 [215/720] Linking static target lib/librte_bpf.a 00:02:21.735 [216/720] Linking target lib/librte_mbuf.so.25.0 00:02:21.995 [217/720] Generating symbol file lib/librte_mbuf.so.25.0.p/librte_mbuf.so.25.0.symbols 00:02:21.995 [218/720] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:21.995 [219/720] Linking target lib/librte_net.so.25.0 00:02:21.995 [220/720] Linking target lib/librte_bbdev.so.25.0 00:02:21.995 [221/720] Linking static target lib/librte_compressdev.a 00:02:21.995 [222/720] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:21.995 [223/720] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.995 [224/720] Linking static target lib/librte_acl.a 00:02:21.995 [225/720] Generating symbol file lib/librte_net.so.25.0.p/librte_net.so.25.0.symbols 00:02:21.995 [226/720] Linking target lib/librte_cmdline.so.25.0 00:02:21.995 [227/720] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:21.995 [228/720] Linking target lib/librte_hash.so.25.0 00:02:22.256 [229/720] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:22.256 [230/720] Generating symbol file lib/librte_hash.so.25.0.p/librte_hash.so.25.0.symbols 00:02:22.256 [231/720] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.256 [232/720] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:22.256 [233/720] Linking target lib/librte_acl.so.25.0 00:02:22.256 [234/720] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:22.256 [235/720] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.516 [236/720] Linking target lib/librte_compressdev.so.25.0 00:02:22.516 [237/720] Generating symbol file lib/librte_acl.so.25.0.p/librte_acl.so.25.0.symbols 00:02:22.516 [238/720] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:22.516 [239/720] Linking static target lib/librte_distributor.a 00:02:22.516 [240/720] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:02:22.516 [241/720] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.776 [242/720] Linking target lib/librte_distributor.so.25.0 00:02:22.776 [243/720] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:22.776 [244/720] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:22.776 [245/720] Linking static target lib/librte_dmadev.a 00:02:23.035 [246/720] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:23.035 [247/720] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.035 [248/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:23.035 [249/720] Linking target lib/librte_dmadev.so.25.0 00:02:23.295 [250/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_dma_adapter.c.o 00:02:23.295 [251/720] Generating symbol file lib/librte_dmadev.so.25.0.p/librte_dmadev.so.25.0.symbols 00:02:23.295 [252/720] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:23.295 [253/720] Linking static target lib/librte_efd.a 00:02:23.554 [254/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:23.554 [255/720] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.554 [256/720] Linking target lib/librte_efd.so.25.0 00:02:23.554 [257/720] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:23.554 [258/720] Linking static target lib/librte_cryptodev.a 00:02:23.815 [259/720] Compiling C object lib/librte_dispatcher.a.p/dispatcher_rte_dispatcher.c.o 00:02:23.815 [260/720] Linking static target lib/librte_dispatcher.a 00:02:23.815 [261/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:23.815 [262/720] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:23.815 [263/720] Linking static target lib/librte_gpudev.a 00:02:23.815 [264/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:24.075 [265/720] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:24.075 [266/720] Generating lib/dispatcher.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.075 [267/720] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:24.343 [268/720] Compiling C object lib/librte_gro.a.p/gro_gro_tcp6.c.o 00:02:24.343 [269/720] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:24.343 [270/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:24.622 [271/720] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:24.622 [272/720] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.622 [273/720] Linking target lib/librte_gpudev.so.25.0 00:02:24.622 [274/720] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:24.622 [275/720] Linking static target lib/librte_gro.a 00:02:24.622 [276/720] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:24.622 [277/720] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.622 [278/720] Linking target lib/librte_cryptodev.so.25.0 00:02:24.622 [279/720] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:24.622 [280/720] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:24.910 [281/720] Generating symbol file lib/librte_cryptodev.so.25.0.p/librte_cryptodev.so.25.0.symbols 00:02:24.910 [282/720] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.910 [283/720] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:24.910 [284/720] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:24.910 [285/720] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:24.910 [286/720] Linking static target lib/librte_eventdev.a 00:02:24.910 [287/720] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:24.910 [288/720] Linking static target lib/librte_gso.a 00:02:24.910 [289/720] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.186 [290/720] Linking target lib/librte_ethdev.so.25.0 00:02:25.186 [291/720] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.186 [292/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:25.186 [293/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:25.186 [294/720] Generating symbol file lib/librte_ethdev.so.25.0.p/librte_ethdev.so.25.0.symbols 00:02:25.186 [295/720] Linking target lib/librte_metrics.so.25.0 00:02:25.186 [296/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:25.186 [297/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:25.186 [298/720] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:25.186 [299/720] Linking target lib/librte_gro.so.25.0 00:02:25.186 [300/720] Linking target lib/librte_bpf.so.25.0 00:02:25.186 [301/720] Linking static target lib/librte_jobstats.a 00:02:25.186 [302/720] Linking target lib/librte_gso.so.25.0 00:02:25.186 [303/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:25.186 [304/720] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:25.186 [305/720] Linking static target lib/librte_ip_frag.a 00:02:25.186 [306/720] Generating symbol file lib/librte_metrics.so.25.0.p/librte_metrics.so.25.0.symbols 00:02:25.445 [307/720] Generating symbol file lib/librte_bpf.so.25.0.p/librte_bpf.so.25.0.symbols 00:02:25.445 [308/720] Linking target lib/librte_bitratestats.so.25.0 00:02:25.445 [309/720] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.445 [310/720] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.445 [311/720] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:25.445 [312/720] Linking target lib/librte_jobstats.so.25.0 00:02:25.445 [313/720] Linking static target lib/librte_latencystats.a 00:02:25.445 [314/720] Linking target lib/librte_ip_frag.so.25.0 00:02:25.705 [315/720] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:25.705 [316/720] Generating symbol file lib/librte_ip_frag.so.25.0.p/librte_ip_frag.so.25.0.symbols 00:02:25.705 [317/720] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:25.705 [318/720] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.705 [319/720] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:25.705 [320/720] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:25.705 [321/720] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:25.705 [322/720] Linking target lib/librte_latencystats.so.25.0 00:02:25.965 [323/720] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:25.965 [324/720] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:25.965 [325/720] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:25.965 [326/720] Linking static target lib/librte_lpm.a 00:02:26.224 [327/720] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:26.224 [328/720] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:26.224 [329/720] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:02:26.224 [330/720] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:26.224 [331/720] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.224 [332/720] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:26.224 [333/720] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:26.224 [334/720] Linking static target lib/librte_pcapng.a 00:02:26.224 [335/720] Linking target lib/librte_lpm.so.25.0 00:02:26.224 [336/720] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:02:26.484 [337/720] Generating symbol file lib/librte_lpm.so.25.0.p/librte_lpm.so.25.0.symbols 00:02:26.484 [338/720] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:26.484 [339/720] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.484 [340/720] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:02:26.484 [341/720] Linking target lib/librte_pcapng.so.25.0 00:02:26.484 [342/720] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.744 [343/720] Linking target lib/librte_eventdev.so.25.0 00:02:26.744 [344/720] Generating symbol file lib/librte_pcapng.so.25.0.p/librte_pcapng.so.25.0.symbols 00:02:26.744 [345/720] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:26.744 [346/720] Generating symbol file lib/librte_eventdev.so.25.0.p/librte_eventdev.so.25.0.symbols 00:02:26.744 [347/720] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:26.744 [348/720] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev_pmd.c.o 00:02:26.744 [349/720] Linking target lib/librte_dispatcher.so.25.0 00:02:26.744 [350/720] Linking static target lib/librte_power.a 00:02:26.744 [351/720] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils.c.o 00:02:26.744 [352/720] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:26.744 [353/720] Linking static target lib/librte_regexdev.a 00:02:26.744 [354/720] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:27.003 [355/720] Linking static target lib/librte_rawdev.a 00:02:27.004 [356/720] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:27.004 [357/720] Linking static target lib/librte_member.a 00:02:27.004 [358/720] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev.c.o 00:02:27.004 [359/720] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar_bfloat16.c.o 00:02:27.263 [360/720] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar.c.o 00:02:27.263 [361/720] Linking static target lib/librte_mldev.a 00:02:27.263 [362/720] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.263 [363/720] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:27.263 [364/720] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.263 [365/720] Linking target lib/librte_member.so.25.0 00:02:27.263 [366/720] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.263 [367/720] Linking target lib/librte_rawdev.so.25.0 00:02:27.263 [368/720] Linking target lib/librte_power.so.25.0 00:02:27.263 [369/720] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:27.263 [370/720] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:27.263 [371/720] Linking static target lib/librte_reorder.a 00:02:27.263 [372/720] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:27.522 [373/720] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.522 [374/720] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:27.522 [375/720] Linking static target lib/librte_rib.a 00:02:27.522 [376/720] Linking target lib/librte_regexdev.so.25.0 00:02:27.522 [377/720] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:27.522 [378/720] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.522 [379/720] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:27.522 [380/720] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:27.780 [381/720] Linking target lib/librte_reorder.so.25.0 00:02:27.780 [382/720] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:27.780 [383/720] Linking static target lib/librte_stack.a 00:02:27.780 [384/720] Generating symbol file lib/librte_reorder.so.25.0.p/librte_reorder.so.25.0.symbols 00:02:27.780 [385/720] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:27.780 [386/720] Linking static target lib/librte_security.a 00:02:27.780 [387/720] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.780 [388/720] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:27.780 [389/720] Linking target lib/librte_rib.so.25.0 00:02:28.040 [390/720] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.040 [391/720] Linking target lib/librte_stack.so.25.0 00:02:28.040 [392/720] Generating symbol file lib/librte_rib.so.25.0.p/librte_rib.so.25.0.symbols 00:02:28.040 [393/720] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:28.040 [394/720] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:28.299 [395/720] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.299 [396/720] Linking target lib/librte_security.so.25.0 00:02:28.299 [397/720] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:28.299 [398/720] Generating symbol file lib/librte_security.so.25.0.p/librte_security.so.25.0.symbols 00:02:28.299 [399/720] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:28.299 [400/720] Linking static target lib/librte_sched.a 00:02:28.557 [401/720] Generating lib/mldev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.557 [402/720] Linking target lib/librte_mldev.so.25.0 00:02:28.557 [403/720] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:02:28.557 [404/720] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:02:28.817 [405/720] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.817 [406/720] Linking target lib/librte_sched.so.25.0 00:02:28.817 [407/720] Generating symbol file lib/librte_sched.so.25.0.p/librte_sched.so.25.0.symbols 00:02:29.139 [408/720] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:29.139 [409/720] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:29.139 [410/720] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:29.139 [411/720] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:29.139 [412/720] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:29.398 [413/720] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_crypto.c.o 00:02:29.398 [414/720] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_ctrl_pdu.c.o 00:02:29.398 [415/720] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_cnt.c.o 00:02:29.656 [416/720] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:29.656 [417/720] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_reorder.c.o 00:02:29.656 [418/720] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:29.914 [419/720] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:29.914 [420/720] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:29.914 [421/720] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:29.914 [422/720] Compiling C object lib/librte_port.a.p/port_port_log.c.o 00:02:29.914 [423/720] Linking static target lib/librte_ipsec.a 00:02:29.914 [424/720] Compiling C object lib/librte_pdcp.a.p/pdcp_rte_pdcp.c.o 00:02:30.177 [425/720] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.177 [426/720] Linking target lib/librte_ipsec.so.25.0 00:02:30.177 [427/720] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:30.177 [428/720] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:30.177 [429/720] Generating symbol file lib/librte_ipsec.so.25.0.p/librte_ipsec.so.25.0.symbols 00:02:30.437 [430/720] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:30.437 [431/720] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:30.437 [432/720] Linking static target lib/librte_fib.a 00:02:30.437 [433/720] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:30.696 [434/720] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:30.696 [435/720] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:30.696 [436/720] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:30.696 [437/720] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.955 [438/720] Linking target lib/librte_fib.so.25.0 00:02:30.955 [439/720] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_process.c.o 00:02:30.955 [440/720] Linking static target lib/librte_pdcp.a 00:02:30.955 [441/720] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:31.215 [442/720] Generating lib/pdcp.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.215 [443/720] Linking target lib/librte_pdcp.so.25.0 00:02:31.215 [444/720] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:31.215 [445/720] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:31.474 [446/720] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:31.474 [447/720] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:31.474 [448/720] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:31.474 [449/720] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:31.733 [450/720] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:31.733 [451/720] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:31.992 [452/720] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:31.992 [453/720] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:31.992 [454/720] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:31.992 [455/720] Linking static target lib/librte_port.a 00:02:31.992 [456/720] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:31.992 [457/720] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:32.252 [458/720] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:32.252 [459/720] Linking static target lib/librte_pdump.a 00:02:32.252 [460/720] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:32.252 [461/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:32.252 [462/720] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.252 [463/720] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:32.512 [464/720] Linking target lib/librte_port.so.25.0 00:02:32.512 [465/720] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.512 [466/720] Linking target lib/librte_pdump.so.25.0 00:02:32.512 [467/720] Generating symbol file lib/librte_port.so.25.0.p/librte_port.so.25.0.symbols 00:02:32.771 [468/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:32.771 [469/720] Compiling C object lib/librte_table.a.p/table_table_log.c.o 00:02:32.771 [470/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:32.771 [471/720] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:32.771 [472/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:32.771 [473/720] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:33.031 [474/720] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:33.291 [475/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:33.291 [476/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:33.291 [477/720] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:33.291 [478/720] Linking static target lib/librte_table.a 00:02:33.550 [479/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:33.550 [480/720] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:33.807 [481/720] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:33.807 [482/720] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.807 [483/720] Linking target lib/librte_table.so.25.0 00:02:34.065 [484/720] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:34.065 [485/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ipsec.c.o 00:02:34.065 [486/720] Generating symbol file lib/librte_table.so.25.0.p/librte_table.so.25.0.symbols 00:02:34.065 [487/720] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:34.065 [488/720] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:34.324 [489/720] Compiling C object lib/librte_graph.a.p/graph_rte_graph_worker.c.o 00:02:34.324 [490/720] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:34.324 [491/720] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:34.586 [492/720] Compiling C object lib/librte_graph.a.p/graph_graph_pcap.c.o 00:02:34.586 [493/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:34.586 [494/720] Compiling C object lib/librte_graph.a.p/graph_rte_graph_model_mcore_dispatch.c.o 00:02:34.846 [495/720] Linking static target lib/librte_graph.a 00:02:34.846 [496/720] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:34.846 [497/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:34.846 [498/720] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:34.846 [499/720] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:35.105 [500/720] Compiling C object lib/librte_node.a.p/node_ip4_local.c.o 00:02:35.105 [501/720] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.364 [502/720] Linking target lib/librte_graph.so.25.0 00:02:35.365 [503/720] Compiling C object lib/librte_node.a.p/node_ip4_reassembly.c.o 00:02:35.365 [504/720] Generating symbol file lib/librte_graph.so.25.0.p/librte_graph.so.25.0.symbols 00:02:35.365 [505/720] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:35.365 [506/720] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:35.625 [507/720] Compiling C object lib/librte_node.a.p/node_ip6_lookup.c.o 00:02:35.625 [508/720] Compiling C object lib/librte_node.a.p/node_kernel_tx.c.o 00:02:35.625 [509/720] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:35.884 [510/720] Compiling C object lib/librte_node.a.p/node_kernel_rx.c.o 00:02:35.885 [511/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:35.885 [512/720] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:35.885 [513/720] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:36.144 [514/720] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:36.144 [515/720] Compiling C object lib/librte_node.a.p/node_ip6_rewrite.c.o 00:02:36.144 [516/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:36.144 [517/720] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:36.405 [518/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:36.405 [519/720] Compiling C object lib/librte_node.a.p/node_udp4_input.c.o 00:02:36.405 [520/720] Linking static target lib/librte_node.a 00:02:36.405 [521/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:36.405 [522/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:36.664 [523/720] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:36.664 [524/720] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:36.664 [525/720] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.664 [526/720] Linking target lib/librte_node.so.25.0 00:02:36.664 [527/720] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:36.664 [528/720] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:36.664 [529/720] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:36.664 [530/720] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:36.664 [531/720] Linking static target drivers/librte_bus_vdev.a 00:02:36.924 [532/720] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:36.924 [533/720] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:36.924 [534/720] Linking static target drivers/librte_bus_pci.a 00:02:36.924 [535/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:36.924 [536/720] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.924 [537/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:36.924 [538/720] Compiling C object drivers/librte_bus_pci.so.25.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:36.924 [539/720] Compiling C object drivers/librte_bus_vdev.so.25.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:36.924 [540/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:36.924 [541/720] Linking target drivers/librte_bus_vdev.so.25.0 00:02:37.182 [542/720] Generating symbol file drivers/librte_bus_vdev.so.25.0.p/librte_bus_vdev.so.25.0.symbols 00:02:37.182 [543/720] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:37.182 [544/720] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:37.441 [545/720] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:37.441 [546/720] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:37.441 [547/720] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:37.441 [548/720] Linking static target drivers/librte_mempool_ring.a 00:02:37.441 [549/720] Compiling C object drivers/librte_mempool_ring.so.25.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:37.441 [550/720] Linking target drivers/librte_bus_pci.so.25.0 00:02:37.441 [551/720] Linking target drivers/librte_mempool_ring.so.25.0 00:02:37.441 [552/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:02:37.441 [553/720] Generating symbol file drivers/librte_bus_pci.so.25.0.p/librte_bus_pci.so.25.0.symbols 00:02:37.700 [554/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:02:37.961 [555/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:02:37.961 [556/720] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:02:37.961 [557/720] Linking static target drivers/net/i40e/base/libi40e_base.a 00:02:38.547 [558/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:02:38.807 [559/720] Compiling C object drivers/net/i40e/libi40e_avx2_lib.a.p/i40e_rxtx_vec_avx2.c.o 00:02:38.807 [560/720] Linking static target drivers/net/i40e/libi40e_avx2_lib.a 00:02:39.067 [561/720] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:02:39.067 [562/720] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:02:39.067 [563/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:02:39.067 [564/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:02:39.067 [565/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:02:39.327 [566/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:02:39.327 [567/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:02:39.587 [568/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:02:39.587 [569/720] Generating app/graph/commands_hdr with a custom command (wrapped by meson to capture output) 00:02:39.587 [570/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_recycle_mbufs_vec_common.c.o 00:02:39.846 [571/720] Compiling C object app/dpdk-graph.p/graph_cli.c.o 00:02:39.846 [572/720] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:02:40.106 [573/720] Compiling C object app/dpdk-graph.p/graph_conn.c.o 00:02:40.366 [574/720] Compiling C object app/dpdk-graph.p/graph_ethdev_rx.c.o 00:02:40.366 [575/720] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:02:40.626 [576/720] Compiling C object app/dpdk-graph.p/graph_ethdev.c.o 00:02:40.626 [577/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:02:40.626 [578/720] Compiling C object app/dpdk-graph.p/graph_graph.c.o 00:02:40.626 [579/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:02:40.626 [580/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:02:40.626 [581/720] Compiling C object app/dpdk-graph.p/graph_ip4_route.c.o 00:02:40.886 [582/720] Compiling C object app/dpdk-graph.p/graph_ip6_route.c.o 00:02:40.886 [583/720] Compiling C object app/dpdk-graph.p/graph_l2fwd.c.o 00:02:40.886 [584/720] Compiling C object app/dpdk-graph.p/graph_main.c.o 00:02:40.886 [585/720] Compiling C object app/dpdk-graph.p/graph_l3fwd.c.o 00:02:40.886 [586/720] Compiling C object app/dpdk-graph.p/graph_mempool.c.o 00:02:41.146 [587/720] Compiling C object app/dpdk-graph.p/graph_utils.c.o 00:02:41.146 [588/720] Compiling C object app/dpdk-graph.p/graph_neigh.c.o 00:02:41.146 [589/720] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:02:41.146 [590/720] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:02:41.404 [591/720] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:02:41.404 [592/720] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:02:41.664 [593/720] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:02:41.664 [594/720] Linking static target drivers/libtmp_rte_net_i40e.a 00:02:41.664 [595/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:02:41.664 [596/720] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:02:41.924 [597/720] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:02:41.924 [598/720] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:02:41.924 [599/720] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:41.924 [600/720] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:02:41.924 [601/720] Linking static target drivers/librte_net_i40e.a 00:02:41.924 [602/720] Compiling C object drivers/librte_net_i40e.so.25.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:42.184 [603/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:02:42.184 [604/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:02:42.184 [605/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:02:42.443 [606/720] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.443 [607/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:02:42.443 [608/720] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:02:42.443 [609/720] Linking target drivers/librte_net_i40e.so.25.0 00:02:42.702 [610/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:02:42.702 [611/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:02:42.962 [612/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:02:42.962 [613/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:02:42.962 [614/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:02:43.221 [615/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:02:43.222 [616/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:02:43.222 [617/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:02:43.222 [618/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:02:43.222 [619/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:02:43.481 [620/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:02:43.481 [621/720] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:43.481 [622/720] Linking static target lib/librte_vhost.a 00:02:43.481 [623/720] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:02:43.481 [624/720] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_main.c.o 00:02:43.740 [625/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:02:43.740 [626/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:02:43.999 [627/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:02:43.999 [628/720] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_benchmark.c.o 00:02:43.999 [629/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:02:44.258 [630/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:02:44.258 [631/720] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.524 [632/720] Linking target lib/librte_vhost.so.25.0 00:02:44.785 [633/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:02:44.785 [634/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:02:44.785 [635/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:02:44.785 [636/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:02:45.045 [637/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:02:45.045 [638/720] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:02:45.045 [639/720] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:02:45.045 [640/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_test.c.o 00:02:45.304 [641/720] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:02:45.304 [642/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_parser.c.o 00:02:45.304 [643/720] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:02:45.304 [644/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_main.c.o 00:02:45.304 [645/720] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:02:45.304 [646/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_options.c.o 00:02:45.563 [647/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_device_ops.c.o 00:02:45.563 [648/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_common.c.o 00:02:45.563 [649/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_common.c.o 00:02:45.822 [650/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_ops.c.o 00:02:45.822 [651/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_ordered.c.o 00:02:45.822 [652/720] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:02:45.823 [653/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_interleave.c.o 00:02:45.823 [654/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_stats.c.o 00:02:46.082 [655/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:02:46.082 [656/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:02:46.082 [657/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:02:46.342 [658/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:02:46.342 [659/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:02:46.342 [660/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:02:46.342 [661/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:02:46.602 [662/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:02:46.861 [663/720] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:02:46.861 [664/720] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:02:46.861 [665/720] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:02:46.861 [666/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_cman.c.o 00:02:46.861 [667/720] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:02:46.861 [668/720] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_common.c.o 00:02:47.121 [669/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:02:47.121 [670/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:02:47.380 [671/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:02:47.380 [672/720] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:02:47.640 [673/720] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:02:47.640 [674/720] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:02:47.640 [675/720] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:02:47.927 [676/720] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:02:47.927 [677/720] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:02:48.187 [678/720] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:02:48.187 [679/720] Compiling C object app/dpdk-testpmd.p/test-pmd_recycle_mbufs.c.o 00:02:48.187 [680/720] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:02:48.187 [681/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:02:48.447 [682/720] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:02:48.447 [683/720] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:02:48.707 [684/720] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:02:48.967 [685/720] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:02:48.967 [686/720] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:02:49.227 [687/720] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:02:49.227 [688/720] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:02:49.227 [689/720] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:02:49.487 [690/720] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:02:49.487 [691/720] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:02:49.487 [692/720] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:02:49.487 [693/720] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:02:49.747 [694/720] Compiling C object app/dpdk-test-security-perf.p/test_test_security_proto.c.o 00:02:49.747 [695/720] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:02:50.006 [696/720] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:02:50.575 [697/720] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:02:50.844 [698/720] Linking static target lib/librte_pipeline.a 00:02:51.119 [699/720] Linking target app/dpdk-test-bbdev 00:02:51.119 [700/720] Linking target app/dpdk-pdump 00:02:51.119 [701/720] Linking target app/dpdk-graph 00:02:51.119 [702/720] Linking target app/dpdk-proc-info 00:02:51.119 [703/720] Linking target app/dpdk-test-acl 00:02:51.119 [704/720] Linking target app/dpdk-test-crypto-perf 00:02:51.119 [705/720] Linking target app/dpdk-dumpcap 00:02:51.425 [706/720] Linking target app/dpdk-test-compress-perf 00:02:51.425 [707/720] Linking target app/dpdk-test-cmdline 00:02:51.684 [708/720] Linking target app/dpdk-test-fib 00:02:51.685 [709/720] Linking target app/dpdk-test-dma-perf 00:02:51.685 [710/720] Linking target app/dpdk-test-gpudev 00:02:51.685 [711/720] Linking target app/dpdk-test-flow-perf 00:02:51.685 [712/720] Linking target app/dpdk-test-pipeline 00:02:51.685 [713/720] Linking target app/dpdk-test-regex 00:02:51.685 [714/720] Linking target app/dpdk-test-eventdev 00:02:51.685 [715/720] Linking target app/dpdk-testpmd 00:02:51.685 [716/720] Linking target app/dpdk-test-mldev 00:02:51.944 [717/720] Linking target app/dpdk-test-sad 00:02:51.944 [718/720] Linking target app/dpdk-test-security-perf 00:02:57.241 [719/720] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.241 [720/720] Linking target lib/librte_pipeline.so.25.0 00:02:57.241 14:02:01 build_native_dpdk -- common/autobuild_common.sh@194 -- $ uname -s 00:02:57.241 14:02:01 build_native_dpdk -- common/autobuild_common.sh@194 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:02:57.241 14:02:01 build_native_dpdk -- common/autobuild_common.sh@207 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:02:57.241 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:57.241 [0/1] Installing files. 00:02:57.241 Installing subdir /home/vagrant/spdk_repo/dpdk/usertools/telemetry-endpoints to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/telemetry-endpoints 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/usertools/telemetry-endpoints/counters.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/telemetry-endpoints 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/usertools/telemetry-endpoints/cpu.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/telemetry-endpoints 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/usertools/telemetry-endpoints/memory.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/telemetry-endpoints 00:02:57.241 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.241 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:57.242 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.243 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec_sa.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipv6_addr_swap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipv6_addr_swap.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.244 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:57.245 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:57.245 Installing lib/librte_log.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_log.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_kvargs.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_argparse.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_argparse.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_telemetry.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_eal.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_ring.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_rcu.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_mempool.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_mbuf.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_net.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_meter.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_ethdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_pci.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cmdline.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_metrics.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_hash.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_timer.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_acl.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bbdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bitratestats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_bpf.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cfgfile.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_compressdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_cryptodev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_distributor.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_dmadev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_efd.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_eventdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_dispatcher.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_dispatcher.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_gpudev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.245 Installing lib/librte_gro.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_gso.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_ip_frag.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_jobstats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_latencystats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_lpm.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_member.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_pcapng.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_power.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_rawdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_regexdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_mldev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_mldev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_rib.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_reorder.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_sched.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_security.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_stack.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.246 Installing lib/librte_vhost.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_ipsec.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pdcp.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pdcp.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_fib.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_port.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pdump.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_table.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_pipeline.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_graph.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing lib/librte_node.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing drivers/librte_bus_pci.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0 00:02:57.818 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing drivers/librte_bus_vdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0 00:02:57.818 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing drivers/librte_mempool_ring.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0 00:02:57.818 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.818 Installing drivers/librte_net_i40e.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0 00:02:57.818 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-graph to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-dma-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-mldev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/log/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/argparse/rte_argparse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.818 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lock_annotations.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_stdatomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ptr_compress/rte_ptr_compress.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_dtls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.819 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_pdcp_hdr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_dma_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/dispatcher/rte_dispatcher.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.820 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_mcore_dispatch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_rtc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip6_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_udp4_input_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/buildtools/dpdk-cmdline-gen.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-rss-flows.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry-exporter.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:57.821 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:57.821 Installing symlink pointing to librte_log.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so.25 00:02:57.821 Installing symlink pointing to librte_log.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so 00:02:57.821 Installing symlink pointing to librte_kvargs.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.25 00:02:57.821 Installing symlink pointing to librte_kvargs.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:02:57.821 Installing symlink pointing to librte_argparse.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_argparse.so.25 00:02:57.821 Installing symlink pointing to librte_argparse.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_argparse.so 00:02:57.821 Installing symlink pointing to librte_telemetry.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.25 00:02:57.821 Installing symlink pointing to librte_telemetry.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:02:57.821 Installing symlink pointing to librte_eal.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.25 00:02:57.821 Installing symlink pointing to librte_eal.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:02:57.822 Installing symlink pointing to librte_ring.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.25 00:02:57.822 Installing symlink pointing to librte_ring.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:02:57.822 Installing symlink pointing to librte_rcu.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.25 00:02:57.822 Installing symlink pointing to librte_rcu.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:02:57.822 Installing symlink pointing to librte_mempool.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.25 00:02:57.822 Installing symlink pointing to librte_mempool.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:02:57.822 Installing symlink pointing to librte_mbuf.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.25 00:02:57.822 Installing symlink pointing to librte_mbuf.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:02:57.822 Installing symlink pointing to librte_net.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.25 00:02:57.822 Installing symlink pointing to librte_net.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:02:57.822 Installing symlink pointing to librte_meter.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.25 00:02:57.822 Installing symlink pointing to librte_meter.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:02:57.822 Installing symlink pointing to librte_ethdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.25 00:02:57.822 Installing symlink pointing to librte_ethdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:02:57.822 Installing symlink pointing to librte_pci.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.25 00:02:57.822 Installing symlink pointing to librte_pci.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:02:57.822 Installing symlink pointing to librte_cmdline.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.25 00:02:57.822 Installing symlink pointing to librte_cmdline.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:02:57.822 Installing symlink pointing to librte_metrics.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.25 00:02:57.822 Installing symlink pointing to librte_metrics.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:02:57.822 Installing symlink pointing to librte_hash.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.25 00:02:57.822 Installing symlink pointing to librte_hash.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:02:57.822 Installing symlink pointing to librte_timer.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.25 00:02:57.822 Installing symlink pointing to librte_timer.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:02:57.822 Installing symlink pointing to librte_acl.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.25 00:02:57.822 Installing symlink pointing to librte_acl.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:02:57.822 Installing symlink pointing to librte_bbdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.25 00:02:57.822 Installing symlink pointing to librte_bbdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:02:57.822 Installing symlink pointing to librte_bitratestats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.25 00:02:57.822 Installing symlink pointing to librte_bitratestats.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:02:57.822 Installing symlink pointing to librte_bpf.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.25 00:02:57.822 Installing symlink pointing to librte_bpf.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:02:57.822 Installing symlink pointing to librte_cfgfile.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.25 00:02:57.822 Installing symlink pointing to librte_cfgfile.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:02:57.822 Installing symlink pointing to librte_compressdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.25 00:02:57.822 Installing symlink pointing to librte_compressdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:02:57.822 Installing symlink pointing to librte_cryptodev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.25 00:02:57.822 Installing symlink pointing to librte_cryptodev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:02:57.822 Installing symlink pointing to librte_distributor.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.25 00:02:57.822 Installing symlink pointing to librte_distributor.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:02:57.822 Installing symlink pointing to librte_dmadev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.25 00:02:57.822 Installing symlink pointing to librte_dmadev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:02:57.822 Installing symlink pointing to librte_efd.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.25 00:02:57.822 Installing symlink pointing to librte_efd.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:02:57.822 Installing symlink pointing to librte_eventdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.25 00:02:57.822 Installing symlink pointing to librte_eventdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:02:57.822 Installing symlink pointing to librte_dispatcher.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so.25 00:02:57.822 Installing symlink pointing to librte_dispatcher.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so 00:02:57.822 Installing symlink pointing to librte_gpudev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.25 00:02:57.822 Installing symlink pointing to librte_gpudev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:02:57.822 Installing symlink pointing to librte_gro.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.25 00:02:57.822 Installing symlink pointing to librte_gro.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:02:57.822 Installing symlink pointing to librte_gso.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.25 00:02:57.822 Installing symlink pointing to librte_gso.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:02:57.822 Installing symlink pointing to librte_ip_frag.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.25 00:02:57.822 Installing symlink pointing to librte_ip_frag.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:02:57.822 Installing symlink pointing to librte_jobstats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.25 00:02:57.822 Installing symlink pointing to librte_jobstats.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:02:57.822 Installing symlink pointing to librte_latencystats.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.25 00:02:57.822 Installing symlink pointing to librte_latencystats.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:02:57.822 Installing symlink pointing to librte_lpm.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.25 00:02:57.822 Installing symlink pointing to librte_lpm.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:02:57.822 Installing symlink pointing to librte_member.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.25 00:02:57.822 Installing symlink pointing to librte_member.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:02:57.822 Installing symlink pointing to librte_pcapng.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.25 00:02:57.822 './librte_bus_pci.so' -> 'dpdk/pmds-25.0/librte_bus_pci.so' 00:02:57.822 './librte_bus_pci.so.25' -> 'dpdk/pmds-25.0/librte_bus_pci.so.25' 00:02:57.822 './librte_bus_pci.so.25.0' -> 'dpdk/pmds-25.0/librte_bus_pci.so.25.0' 00:02:57.822 './librte_bus_vdev.so' -> 'dpdk/pmds-25.0/librte_bus_vdev.so' 00:02:57.822 './librte_bus_vdev.so.25' -> 'dpdk/pmds-25.0/librte_bus_vdev.so.25' 00:02:57.822 './librte_bus_vdev.so.25.0' -> 'dpdk/pmds-25.0/librte_bus_vdev.so.25.0' 00:02:57.822 './librte_mempool_ring.so' -> 'dpdk/pmds-25.0/librte_mempool_ring.so' 00:02:57.822 './librte_mempool_ring.so.25' -> 'dpdk/pmds-25.0/librte_mempool_ring.so.25' 00:02:57.822 './librte_mempool_ring.so.25.0' -> 'dpdk/pmds-25.0/librte_mempool_ring.so.25.0' 00:02:57.822 './librte_net_i40e.so' -> 'dpdk/pmds-25.0/librte_net_i40e.so' 00:02:57.822 './librte_net_i40e.so.25' -> 'dpdk/pmds-25.0/librte_net_i40e.so.25' 00:02:57.822 './librte_net_i40e.so.25.0' -> 'dpdk/pmds-25.0/librte_net_i40e.so.25.0' 00:02:57.822 Installing symlink pointing to librte_pcapng.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:02:57.822 Installing symlink pointing to librte_power.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.25 00:02:57.822 Installing symlink pointing to librte_power.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:02:57.822 Installing symlink pointing to librte_rawdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.25 00:02:57.822 Installing symlink pointing to librte_rawdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:02:57.822 Installing symlink pointing to librte_regexdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.25 00:02:57.822 Installing symlink pointing to librte_regexdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:02:57.822 Installing symlink pointing to librte_mldev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so.25 00:02:57.822 Installing symlink pointing to librte_mldev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so 00:02:57.822 Installing symlink pointing to librte_rib.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.25 00:02:57.822 Installing symlink pointing to librte_rib.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:02:57.822 Installing symlink pointing to librte_reorder.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.25 00:02:57.822 Installing symlink pointing to librte_reorder.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:02:57.822 Installing symlink pointing to librte_sched.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.25 00:02:57.822 Installing symlink pointing to librte_sched.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:02:57.822 Installing symlink pointing to librte_security.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.25 00:02:57.822 Installing symlink pointing to librte_security.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:02:57.822 Installing symlink pointing to librte_stack.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.25 00:02:57.822 Installing symlink pointing to librte_stack.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:02:57.822 Installing symlink pointing to librte_vhost.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.25 00:02:57.822 Installing symlink pointing to librte_vhost.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:02:57.822 Installing symlink pointing to librte_ipsec.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.25 00:02:57.822 Installing symlink pointing to librte_ipsec.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:02:57.822 Installing symlink pointing to librte_pdcp.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so.25 00:02:57.822 Installing symlink pointing to librte_pdcp.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so 00:02:57.822 Installing symlink pointing to librte_fib.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.25 00:02:57.822 Installing symlink pointing to librte_fib.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:02:57.823 Installing symlink pointing to librte_port.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.25 00:02:57.823 Installing symlink pointing to librte_port.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:02:57.823 Installing symlink pointing to librte_pdump.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.25 00:02:57.823 Installing symlink pointing to librte_pdump.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:02:57.823 Installing symlink pointing to librte_table.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.25 00:02:57.823 Installing symlink pointing to librte_table.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:02:57.823 Installing symlink pointing to librte_pipeline.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.25 00:02:57.823 Installing symlink pointing to librte_pipeline.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:02:57.823 Installing symlink pointing to librte_graph.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.25 00:02:57.823 Installing symlink pointing to librte_graph.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:02:57.823 Installing symlink pointing to librte_node.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.25 00:02:57.823 Installing symlink pointing to librte_node.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:02:57.823 Installing symlink pointing to librte_bus_pci.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_pci.so.25 00:02:57.823 Installing symlink pointing to librte_bus_pci.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_pci.so 00:02:57.823 Installing symlink pointing to librte_bus_vdev.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_vdev.so.25 00:02:57.823 Installing symlink pointing to librte_bus_vdev.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_vdev.so 00:02:57.823 Installing symlink pointing to librte_mempool_ring.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_mempool_ring.so.25 00:02:57.823 Installing symlink pointing to librte_mempool_ring.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_mempool_ring.so 00:02:57.823 Installing symlink pointing to librte_net_i40e.so.25.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_net_i40e.so.25 00:02:57.823 Installing symlink pointing to librte_net_i40e.so.25 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_net_i40e.so 00:02:57.823 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-25.0' 00:02:57.823 14:02:02 build_native_dpdk -- common/autobuild_common.sh@213 -- $ cat 00:02:57.823 14:02:02 build_native_dpdk -- common/autobuild_common.sh@218 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:57.823 00:02:57.823 real 0m52.111s 00:02:57.823 user 5m17.744s 00:02:57.823 sys 0m59.171s 00:02:57.823 14:02:02 build_native_dpdk -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:57.823 14:02:02 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:02:57.823 ************************************ 00:02:57.823 END TEST build_native_dpdk 00:02:57.823 ************************************ 00:02:57.823 14:02:02 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:02:57.823 14:02:02 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:02:57.823 14:02:02 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:02:58.083 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:02:58.083 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:02:58.083 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:02:58.083 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:02:58.650 Using 'verbs' RDMA provider 00:03:14.922 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:33.027 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:33.027 Creating mk/config.mk...done. 00:03:33.027 Creating mk/cc.flags.mk...done. 00:03:33.027 Type 'make' to build. 00:03:33.027 14:02:36 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:33.027 14:02:36 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:03:33.027 14:02:36 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:03:33.027 14:02:36 -- common/autotest_common.sh@10 -- $ set +x 00:03:33.027 ************************************ 00:03:33.027 START TEST make 00:03:33.027 ************************************ 00:03:33.027 14:02:36 make -- common/autotest_common.sh@1125 -- $ make -j10 00:03:33.027 make[1]: Nothing to be done for 'all'. 00:04:11.753 CC lib/log/log.o 00:04:11.753 CC lib/log/log_flags.o 00:04:11.753 CC lib/log/log_deprecated.o 00:04:11.753 CC lib/ut/ut.o 00:04:11.753 CC lib/ut_mock/mock.o 00:04:11.753 LIB libspdk_log.a 00:04:11.753 LIB libspdk_ut_mock.a 00:04:11.754 LIB libspdk_ut.a 00:04:11.754 SO libspdk_ut.so.2.0 00:04:12.014 SO libspdk_log.so.7.0 00:04:12.014 SO libspdk_ut_mock.so.6.0 00:04:12.014 SYMLINK libspdk_log.so 00:04:12.014 SYMLINK libspdk_ut_mock.so 00:04:12.014 SYMLINK libspdk_ut.so 00:04:12.273 CC lib/ioat/ioat.o 00:04:12.273 CC lib/dma/dma.o 00:04:12.273 CC lib/util/bit_array.o 00:04:12.273 CC lib/util/base64.o 00:04:12.273 CC lib/util/crc32.o 00:04:12.273 CC lib/util/cpuset.o 00:04:12.273 CC lib/util/crc16.o 00:04:12.273 CC lib/util/crc32c.o 00:04:12.273 CXX lib/trace_parser/trace.o 00:04:12.273 CC lib/vfio_user/host/vfio_user_pci.o 00:04:12.273 CC lib/util/crc32_ieee.o 00:04:12.273 CC lib/util/crc64.o 00:04:12.273 CC lib/util/dif.o 00:04:12.543 CC lib/util/fd.o 00:04:12.543 LIB libspdk_dma.a 00:04:12.543 CC lib/util/fd_group.o 00:04:12.543 CC lib/vfio_user/host/vfio_user.o 00:04:12.543 SO libspdk_dma.so.5.0 00:04:12.543 LIB libspdk_ioat.a 00:04:12.543 CC lib/util/file.o 00:04:12.543 CC lib/util/hexlify.o 00:04:12.543 SYMLINK libspdk_dma.so 00:04:12.543 CC lib/util/iov.o 00:04:12.543 SO libspdk_ioat.so.7.0 00:04:12.543 CC lib/util/math.o 00:04:12.543 SYMLINK libspdk_ioat.so 00:04:12.543 CC lib/util/net.o 00:04:12.543 CC lib/util/pipe.o 00:04:12.543 CC lib/util/strerror_tls.o 00:04:12.543 LIB libspdk_vfio_user.a 00:04:12.543 CC lib/util/string.o 00:04:12.543 CC lib/util/uuid.o 00:04:12.543 SO libspdk_vfio_user.so.5.0 00:04:12.823 CC lib/util/xor.o 00:04:12.823 SYMLINK libspdk_vfio_user.so 00:04:12.823 CC lib/util/zipf.o 00:04:12.823 CC lib/util/md5.o 00:04:13.092 LIB libspdk_util.a 00:04:13.092 SO libspdk_util.so.10.0 00:04:13.092 LIB libspdk_trace_parser.a 00:04:13.092 SYMLINK libspdk_util.so 00:04:13.356 SO libspdk_trace_parser.so.6.0 00:04:13.356 SYMLINK libspdk_trace_parser.so 00:04:13.356 CC lib/conf/conf.o 00:04:13.356 CC lib/rdma_utils/rdma_utils.o 00:04:13.356 CC lib/json/json_parse.o 00:04:13.356 CC lib/json/json_util.o 00:04:13.356 CC lib/env_dpdk/env.o 00:04:13.356 CC lib/env_dpdk/memory.o 00:04:13.356 CC lib/json/json_write.o 00:04:13.356 CC lib/idxd/idxd.o 00:04:13.356 CC lib/vmd/vmd.o 00:04:13.356 CC lib/rdma_provider/common.o 00:04:13.616 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:13.616 LIB libspdk_conf.a 00:04:13.616 CC lib/idxd/idxd_user.o 00:04:13.616 SO libspdk_conf.so.6.0 00:04:13.616 CC lib/vmd/led.o 00:04:13.616 LIB libspdk_rdma_utils.a 00:04:13.616 LIB libspdk_json.a 00:04:13.616 SYMLINK libspdk_conf.so 00:04:13.616 SO libspdk_rdma_utils.so.1.0 00:04:13.616 CC lib/env_dpdk/pci.o 00:04:13.616 SO libspdk_json.so.6.0 00:04:13.616 LIB libspdk_rdma_provider.a 00:04:13.616 SYMLINK libspdk_rdma_utils.so 00:04:13.616 CC lib/env_dpdk/init.o 00:04:13.616 SO libspdk_rdma_provider.so.6.0 00:04:13.616 SYMLINK libspdk_json.so 00:04:13.616 CC lib/env_dpdk/threads.o 00:04:13.616 CC lib/env_dpdk/pci_ioat.o 00:04:13.876 SYMLINK libspdk_rdma_provider.so 00:04:13.876 CC lib/env_dpdk/pci_virtio.o 00:04:13.876 CC lib/idxd/idxd_kernel.o 00:04:13.876 CC lib/env_dpdk/pci_vmd.o 00:04:13.876 CC lib/env_dpdk/pci_idxd.o 00:04:13.876 CC lib/jsonrpc/jsonrpc_server.o 00:04:13.876 CC lib/env_dpdk/pci_event.o 00:04:13.876 CC lib/env_dpdk/sigbus_handler.o 00:04:13.876 CC lib/env_dpdk/pci_dpdk.o 00:04:13.876 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:13.876 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:13.876 LIB libspdk_idxd.a 00:04:14.137 SO libspdk_idxd.so.12.1 00:04:14.137 LIB libspdk_vmd.a 00:04:14.137 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:14.137 SO libspdk_vmd.so.6.0 00:04:14.137 CC lib/jsonrpc/jsonrpc_client.o 00:04:14.137 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:14.137 SYMLINK libspdk_idxd.so 00:04:14.137 SYMLINK libspdk_vmd.so 00:04:14.397 LIB libspdk_jsonrpc.a 00:04:14.397 SO libspdk_jsonrpc.so.6.0 00:04:14.397 SYMLINK libspdk_jsonrpc.so 00:04:14.966 CC lib/rpc/rpc.o 00:04:14.966 LIB libspdk_env_dpdk.a 00:04:14.966 LIB libspdk_rpc.a 00:04:14.966 SO libspdk_env_dpdk.so.15.0 00:04:14.966 SO libspdk_rpc.so.6.0 00:04:15.225 SYMLINK libspdk_rpc.so 00:04:15.225 SYMLINK libspdk_env_dpdk.so 00:04:15.484 CC lib/notify/notify.o 00:04:15.484 CC lib/notify/notify_rpc.o 00:04:15.484 CC lib/keyring/keyring_rpc.o 00:04:15.484 CC lib/keyring/keyring.o 00:04:15.484 CC lib/trace/trace.o 00:04:15.484 CC lib/trace/trace_flags.o 00:04:15.484 CC lib/trace/trace_rpc.o 00:04:15.748 LIB libspdk_notify.a 00:04:15.748 SO libspdk_notify.so.6.0 00:04:15.748 LIB libspdk_keyring.a 00:04:15.748 SO libspdk_keyring.so.2.0 00:04:15.748 SYMLINK libspdk_notify.so 00:04:15.748 LIB libspdk_trace.a 00:04:15.748 SYMLINK libspdk_keyring.so 00:04:15.748 SO libspdk_trace.so.11.0 00:04:16.008 SYMLINK libspdk_trace.so 00:04:16.267 CC lib/thread/thread.o 00:04:16.267 CC lib/thread/iobuf.o 00:04:16.267 CC lib/sock/sock.o 00:04:16.267 CC lib/sock/sock_rpc.o 00:04:16.837 LIB libspdk_sock.a 00:04:16.837 SO libspdk_sock.so.10.0 00:04:16.837 SYMLINK libspdk_sock.so 00:04:17.406 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:17.406 CC lib/nvme/nvme_ctrlr.o 00:04:17.406 CC lib/nvme/nvme_fabric.o 00:04:17.406 CC lib/nvme/nvme_ns_cmd.o 00:04:17.406 CC lib/nvme/nvme_ns.o 00:04:17.406 CC lib/nvme/nvme_pcie_common.o 00:04:17.406 CC lib/nvme/nvme_qpair.o 00:04:17.406 CC lib/nvme/nvme_pcie.o 00:04:17.406 CC lib/nvme/nvme.o 00:04:17.974 LIB libspdk_thread.a 00:04:17.974 CC lib/nvme/nvme_quirks.o 00:04:17.974 SO libspdk_thread.so.10.1 00:04:17.974 CC lib/nvme/nvme_transport.o 00:04:17.974 SYMLINK libspdk_thread.so 00:04:17.975 CC lib/nvme/nvme_discovery.o 00:04:17.975 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:17.975 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:17.975 CC lib/nvme/nvme_tcp.o 00:04:17.975 CC lib/nvme/nvme_opal.o 00:04:18.233 CC lib/nvme/nvme_io_msg.o 00:04:18.233 CC lib/nvme/nvme_poll_group.o 00:04:18.233 CC lib/nvme/nvme_zns.o 00:04:18.492 CC lib/nvme/nvme_stubs.o 00:04:18.492 CC lib/accel/accel.o 00:04:18.751 CC lib/blob/blobstore.o 00:04:18.751 CC lib/nvme/nvme_auth.o 00:04:18.751 CC lib/init/json_config.o 00:04:18.751 CC lib/init/subsystem.o 00:04:18.751 CC lib/nvme/nvme_cuse.o 00:04:18.751 CC lib/nvme/nvme_rdma.o 00:04:18.751 CC lib/blob/request.o 00:04:19.009 CC lib/accel/accel_rpc.o 00:04:19.009 CC lib/init/subsystem_rpc.o 00:04:19.009 CC lib/init/rpc.o 00:04:19.009 CC lib/blob/zeroes.o 00:04:19.268 CC lib/accel/accel_sw.o 00:04:19.268 LIB libspdk_init.a 00:04:19.268 SO libspdk_init.so.6.0 00:04:19.268 SYMLINK libspdk_init.so 00:04:19.268 CC lib/virtio/virtio.o 00:04:19.528 CC lib/fsdev/fsdev.o 00:04:19.528 CC lib/event/app.o 00:04:19.528 CC lib/fsdev/fsdev_io.o 00:04:19.528 CC lib/event/reactor.o 00:04:19.528 CC lib/blob/blob_bs_dev.o 00:04:19.528 LIB libspdk_accel.a 00:04:19.528 CC lib/event/log_rpc.o 00:04:19.787 CC lib/virtio/virtio_vhost_user.o 00:04:19.787 SO libspdk_accel.so.16.0 00:04:19.787 SYMLINK libspdk_accel.so 00:04:19.787 CC lib/fsdev/fsdev_rpc.o 00:04:19.787 CC lib/event/app_rpc.o 00:04:19.787 CC lib/event/scheduler_static.o 00:04:19.787 CC lib/virtio/virtio_vfio_user.o 00:04:19.787 CC lib/virtio/virtio_pci.o 00:04:20.047 CC lib/bdev/bdev.o 00:04:20.047 CC lib/bdev/bdev_rpc.o 00:04:20.047 CC lib/bdev/bdev_zone.o 00:04:20.047 CC lib/bdev/part.o 00:04:20.047 LIB libspdk_fsdev.a 00:04:20.047 SO libspdk_fsdev.so.1.0 00:04:20.047 CC lib/bdev/scsi_nvme.o 00:04:20.047 LIB libspdk_event.a 00:04:20.047 SYMLINK libspdk_fsdev.so 00:04:20.047 SO libspdk_event.so.14.0 00:04:20.047 LIB libspdk_nvme.a 00:04:20.047 LIB libspdk_virtio.a 00:04:20.047 SYMLINK libspdk_event.so 00:04:20.305 SO libspdk_virtio.so.7.0 00:04:20.305 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:20.305 SYMLINK libspdk_virtio.so 00:04:20.305 SO libspdk_nvme.so.14.0 00:04:20.563 SYMLINK libspdk_nvme.so 00:04:20.822 LIB libspdk_fuse_dispatcher.a 00:04:20.822 SO libspdk_fuse_dispatcher.so.1.0 00:04:21.081 SYMLINK libspdk_fuse_dispatcher.so 00:04:22.021 LIB libspdk_blob.a 00:04:22.021 SO libspdk_blob.so.11.0 00:04:22.021 SYMLINK libspdk_blob.so 00:04:22.589 CC lib/lvol/lvol.o 00:04:22.589 CC lib/blobfs/blobfs.o 00:04:22.589 CC lib/blobfs/tree.o 00:04:22.589 LIB libspdk_bdev.a 00:04:22.589 SO libspdk_bdev.so.16.0 00:04:22.848 SYMLINK libspdk_bdev.so 00:04:23.107 CC lib/ublk/ublk.o 00:04:23.107 CC lib/ublk/ublk_rpc.o 00:04:23.107 CC lib/nbd/nbd.o 00:04:23.107 CC lib/nbd/nbd_rpc.o 00:04:23.107 CC lib/scsi/dev.o 00:04:23.107 CC lib/scsi/lun.o 00:04:23.107 CC lib/nvmf/ctrlr.o 00:04:23.107 CC lib/ftl/ftl_core.o 00:04:23.107 CC lib/ftl/ftl_init.o 00:04:23.107 CC lib/scsi/port.o 00:04:23.367 CC lib/ftl/ftl_layout.o 00:04:23.367 CC lib/scsi/scsi.o 00:04:23.367 LIB libspdk_blobfs.a 00:04:23.367 CC lib/ftl/ftl_debug.o 00:04:23.367 CC lib/ftl/ftl_io.o 00:04:23.367 SO libspdk_blobfs.so.10.0 00:04:23.367 LIB libspdk_lvol.a 00:04:23.367 SYMLINK libspdk_blobfs.so 00:04:23.367 CC lib/ftl/ftl_sb.o 00:04:23.367 LIB libspdk_nbd.a 00:04:23.367 CC lib/ftl/ftl_l2p.o 00:04:23.367 CC lib/scsi/scsi_bdev.o 00:04:23.367 SO libspdk_lvol.so.10.0 00:04:23.367 SO libspdk_nbd.so.7.0 00:04:23.626 SYMLINK libspdk_lvol.so 00:04:23.626 SYMLINK libspdk_nbd.so 00:04:23.626 CC lib/scsi/scsi_pr.o 00:04:23.626 CC lib/scsi/scsi_rpc.o 00:04:23.626 CC lib/nvmf/ctrlr_discovery.o 00:04:23.626 CC lib/ftl/ftl_l2p_flat.o 00:04:23.626 CC lib/nvmf/ctrlr_bdev.o 00:04:23.626 CC lib/ftl/ftl_nv_cache.o 00:04:23.626 CC lib/ftl/ftl_band.o 00:04:23.626 LIB libspdk_ublk.a 00:04:23.626 CC lib/ftl/ftl_band_ops.o 00:04:23.626 SO libspdk_ublk.so.3.0 00:04:23.626 SYMLINK libspdk_ublk.so 00:04:23.626 CC lib/ftl/ftl_writer.o 00:04:23.626 CC lib/ftl/ftl_rq.o 00:04:23.885 CC lib/ftl/ftl_reloc.o 00:04:23.885 CC lib/ftl/ftl_l2p_cache.o 00:04:23.885 CC lib/scsi/task.o 00:04:23.885 CC lib/ftl/ftl_p2l.o 00:04:23.885 CC lib/ftl/ftl_p2l_log.o 00:04:24.144 CC lib/ftl/mngt/ftl_mngt.o 00:04:24.144 CC lib/nvmf/subsystem.o 00:04:24.144 LIB libspdk_scsi.a 00:04:24.144 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:24.144 SO libspdk_scsi.so.9.0 00:04:24.403 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:24.403 SYMLINK libspdk_scsi.so 00:04:24.403 CC lib/nvmf/nvmf.o 00:04:24.403 CC lib/nvmf/nvmf_rpc.o 00:04:24.403 CC lib/nvmf/transport.o 00:04:24.403 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:24.403 CC lib/iscsi/conn.o 00:04:24.403 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:24.403 CC lib/vhost/vhost.o 00:04:24.663 CC lib/iscsi/init_grp.o 00:04:24.663 CC lib/iscsi/iscsi.o 00:04:24.663 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:24.922 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:24.922 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:24.922 CC lib/vhost/vhost_rpc.o 00:04:24.922 CC lib/iscsi/param.o 00:04:24.922 CC lib/iscsi/portal_grp.o 00:04:25.181 CC lib/iscsi/tgt_node.o 00:04:25.181 CC lib/iscsi/iscsi_subsystem.o 00:04:25.181 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:25.181 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:25.181 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:25.440 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:25.440 CC lib/iscsi/iscsi_rpc.o 00:04:25.440 CC lib/iscsi/task.o 00:04:25.440 CC lib/vhost/vhost_scsi.o 00:04:25.440 CC lib/vhost/vhost_blk.o 00:04:25.440 CC lib/nvmf/tcp.o 00:04:25.440 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:25.440 CC lib/nvmf/stubs.o 00:04:25.440 CC lib/nvmf/mdns_server.o 00:04:25.440 CC lib/vhost/rte_vhost_user.o 00:04:25.700 CC lib/ftl/utils/ftl_conf.o 00:04:25.700 CC lib/ftl/utils/ftl_md.o 00:04:25.700 CC lib/ftl/utils/ftl_mempool.o 00:04:25.959 CC lib/ftl/utils/ftl_bitmap.o 00:04:25.959 CC lib/ftl/utils/ftl_property.o 00:04:25.959 CC lib/nvmf/rdma.o 00:04:25.959 CC lib/nvmf/auth.o 00:04:25.959 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:25.959 LIB libspdk_iscsi.a 00:04:26.218 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:26.218 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:26.218 SO libspdk_iscsi.so.8.0 00:04:26.218 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:26.218 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:26.218 SYMLINK libspdk_iscsi.so 00:04:26.218 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:26.477 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:26.477 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:26.477 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:26.477 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:26.477 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:26.477 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:26.477 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:26.477 LIB libspdk_vhost.a 00:04:26.736 CC lib/ftl/base/ftl_base_dev.o 00:04:26.736 CC lib/ftl/base/ftl_base_bdev.o 00:04:26.736 SO libspdk_vhost.so.8.0 00:04:26.736 CC lib/ftl/ftl_trace.o 00:04:26.736 SYMLINK libspdk_vhost.so 00:04:26.995 LIB libspdk_ftl.a 00:04:27.254 SO libspdk_ftl.so.9.0 00:04:27.513 SYMLINK libspdk_ftl.so 00:04:28.082 LIB libspdk_nvmf.a 00:04:28.342 SO libspdk_nvmf.so.19.0 00:04:28.601 SYMLINK libspdk_nvmf.so 00:04:28.860 CC module/env_dpdk/env_dpdk_rpc.o 00:04:28.860 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:28.860 CC module/accel/error/accel_error.o 00:04:28.860 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:28.860 CC module/sock/posix/posix.o 00:04:28.860 CC module/keyring/linux/keyring.o 00:04:28.860 CC module/fsdev/aio/fsdev_aio.o 00:04:28.860 CC module/scheduler/gscheduler/gscheduler.o 00:04:28.860 CC module/keyring/file/keyring.o 00:04:28.860 CC module/blob/bdev/blob_bdev.o 00:04:28.860 LIB libspdk_env_dpdk_rpc.a 00:04:29.119 SO libspdk_env_dpdk_rpc.so.6.0 00:04:29.119 SYMLINK libspdk_env_dpdk_rpc.so 00:04:29.119 CC module/keyring/file/keyring_rpc.o 00:04:29.119 CC module/keyring/linux/keyring_rpc.o 00:04:29.119 LIB libspdk_scheduler_dpdk_governor.a 00:04:29.119 CC module/accel/error/accel_error_rpc.o 00:04:29.119 LIB libspdk_scheduler_gscheduler.a 00:04:29.120 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:29.120 SO libspdk_scheduler_gscheduler.so.4.0 00:04:29.120 LIB libspdk_scheduler_dynamic.a 00:04:29.120 SO libspdk_scheduler_dynamic.so.4.0 00:04:29.120 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:29.120 SYMLINK libspdk_scheduler_gscheduler.so 00:04:29.120 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:29.120 LIB libspdk_keyring_linux.a 00:04:29.120 LIB libspdk_keyring_file.a 00:04:29.120 SYMLINK libspdk_scheduler_dynamic.so 00:04:29.120 CC module/fsdev/aio/linux_aio_mgr.o 00:04:29.120 SO libspdk_keyring_linux.so.1.0 00:04:29.120 SO libspdk_keyring_file.so.2.0 00:04:29.120 LIB libspdk_accel_error.a 00:04:29.120 LIB libspdk_blob_bdev.a 00:04:29.120 SO libspdk_blob_bdev.so.11.0 00:04:29.120 SO libspdk_accel_error.so.2.0 00:04:29.379 CC module/accel/ioat/accel_ioat.o 00:04:29.379 SYMLINK libspdk_keyring_linux.so 00:04:29.379 SYMLINK libspdk_keyring_file.so 00:04:29.379 SYMLINK libspdk_blob_bdev.so 00:04:29.379 SYMLINK libspdk_accel_error.so 00:04:29.379 CC module/accel/ioat/accel_ioat_rpc.o 00:04:29.379 CC module/accel/dsa/accel_dsa.o 00:04:29.379 CC module/accel/dsa/accel_dsa_rpc.o 00:04:29.379 CC module/accel/iaa/accel_iaa.o 00:04:29.379 LIB libspdk_accel_ioat.a 00:04:29.379 CC module/accel/iaa/accel_iaa_rpc.o 00:04:29.379 SO libspdk_accel_ioat.so.6.0 00:04:29.638 CC module/bdev/error/vbdev_error.o 00:04:29.638 CC module/blobfs/bdev/blobfs_bdev.o 00:04:29.638 SYMLINK libspdk_accel_ioat.so 00:04:29.638 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:29.638 CC module/bdev/delay/vbdev_delay.o 00:04:29.638 CC module/bdev/gpt/gpt.o 00:04:29.638 CC module/bdev/gpt/vbdev_gpt.o 00:04:29.638 LIB libspdk_fsdev_aio.a 00:04:29.638 LIB libspdk_accel_dsa.a 00:04:29.638 LIB libspdk_accel_iaa.a 00:04:29.638 SO libspdk_fsdev_aio.so.1.0 00:04:29.638 SO libspdk_accel_dsa.so.5.0 00:04:29.638 SO libspdk_accel_iaa.so.3.0 00:04:29.638 LIB libspdk_sock_posix.a 00:04:29.638 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:29.638 LIB libspdk_blobfs_bdev.a 00:04:29.638 SYMLINK libspdk_fsdev_aio.so 00:04:29.638 SO libspdk_sock_posix.so.6.0 00:04:29.638 SYMLINK libspdk_accel_iaa.so 00:04:29.638 SYMLINK libspdk_accel_dsa.so 00:04:29.638 SO libspdk_blobfs_bdev.so.6.0 00:04:29.898 CC module/bdev/error/vbdev_error_rpc.o 00:04:29.898 SYMLINK libspdk_blobfs_bdev.so 00:04:29.898 SYMLINK libspdk_sock_posix.so 00:04:29.898 LIB libspdk_bdev_gpt.a 00:04:29.898 CC module/bdev/lvol/vbdev_lvol.o 00:04:29.898 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:29.898 SO libspdk_bdev_gpt.so.6.0 00:04:29.898 CC module/bdev/malloc/bdev_malloc.o 00:04:29.898 CC module/bdev/null/bdev_null.o 00:04:29.898 LIB libspdk_bdev_delay.a 00:04:29.898 CC module/bdev/nvme/bdev_nvme.o 00:04:29.898 SO libspdk_bdev_delay.so.6.0 00:04:29.898 LIB libspdk_bdev_error.a 00:04:29.898 SYMLINK libspdk_bdev_gpt.so 00:04:29.898 CC module/bdev/passthru/vbdev_passthru.o 00:04:29.898 CC module/bdev/raid/bdev_raid.o 00:04:29.898 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:29.898 SO libspdk_bdev_error.so.6.0 00:04:29.898 SYMLINK libspdk_bdev_delay.so 00:04:29.898 CC module/bdev/raid/bdev_raid_rpc.o 00:04:30.158 SYMLINK libspdk_bdev_error.so 00:04:30.158 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:30.158 CC module/bdev/null/bdev_null_rpc.o 00:04:30.158 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:30.158 CC module/bdev/split/vbdev_split.o 00:04:30.158 CC module/bdev/split/vbdev_split_rpc.o 00:04:30.158 LIB libspdk_bdev_passthru.a 00:04:30.418 SO libspdk_bdev_passthru.so.6.0 00:04:30.418 LIB libspdk_bdev_null.a 00:04:30.418 CC module/bdev/nvme/nvme_rpc.o 00:04:30.418 SO libspdk_bdev_null.so.6.0 00:04:30.418 SYMLINK libspdk_bdev_passthru.so 00:04:30.418 CC module/bdev/raid/bdev_raid_sb.o 00:04:30.418 LIB libspdk_bdev_malloc.a 00:04:30.418 SO libspdk_bdev_malloc.so.6.0 00:04:30.418 LIB libspdk_bdev_lvol.a 00:04:30.418 SYMLINK libspdk_bdev_null.so 00:04:30.418 CC module/bdev/nvme/bdev_mdns_client.o 00:04:30.418 SO libspdk_bdev_lvol.so.6.0 00:04:30.418 SYMLINK libspdk_bdev_malloc.so 00:04:30.418 LIB libspdk_bdev_split.a 00:04:30.418 SO libspdk_bdev_split.so.6.0 00:04:30.418 SYMLINK libspdk_bdev_lvol.so 00:04:30.677 CC module/bdev/nvme/vbdev_opal.o 00:04:30.677 SYMLINK libspdk_bdev_split.so 00:04:30.677 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:30.677 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:30.677 CC module/bdev/aio/bdev_aio.o 00:04:30.677 CC module/bdev/aio/bdev_aio_rpc.o 00:04:30.677 CC module/bdev/ftl/bdev_ftl.o 00:04:30.677 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:30.677 CC module/bdev/iscsi/bdev_iscsi.o 00:04:30.677 CC module/bdev/raid/raid0.o 00:04:30.677 CC module/bdev/raid/raid1.o 00:04:30.677 CC module/bdev/raid/concat.o 00:04:30.937 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:30.937 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:30.937 LIB libspdk_bdev_ftl.a 00:04:30.937 LIB libspdk_bdev_aio.a 00:04:30.938 SO libspdk_bdev_ftl.so.6.0 00:04:30.938 SO libspdk_bdev_aio.so.6.0 00:04:30.938 SYMLINK libspdk_bdev_aio.so 00:04:30.938 SYMLINK libspdk_bdev_ftl.so 00:04:30.938 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:30.938 CC module/bdev/raid/raid5f.o 00:04:30.938 LIB libspdk_bdev_iscsi.a 00:04:30.938 LIB libspdk_bdev_zone_block.a 00:04:31.252 SO libspdk_bdev_iscsi.so.6.0 00:04:31.252 SO libspdk_bdev_zone_block.so.6.0 00:04:31.252 SYMLINK libspdk_bdev_iscsi.so 00:04:31.252 SYMLINK libspdk_bdev_zone_block.so 00:04:31.252 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:31.252 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:31.252 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:31.524 LIB libspdk_bdev_raid.a 00:04:31.784 SO libspdk_bdev_raid.so.6.0 00:04:31.784 LIB libspdk_bdev_virtio.a 00:04:31.784 SYMLINK libspdk_bdev_raid.so 00:04:31.784 SO libspdk_bdev_virtio.so.6.0 00:04:31.784 SYMLINK libspdk_bdev_virtio.so 00:04:32.723 LIB libspdk_bdev_nvme.a 00:04:32.723 SO libspdk_bdev_nvme.so.7.0 00:04:32.723 SYMLINK libspdk_bdev_nvme.so 00:04:33.292 CC module/event/subsystems/keyring/keyring.o 00:04:33.292 CC module/event/subsystems/vmd/vmd.o 00:04:33.292 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:33.292 CC module/event/subsystems/iobuf/iobuf.o 00:04:33.292 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:33.292 CC module/event/subsystems/fsdev/fsdev.o 00:04:33.292 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:33.292 CC module/event/subsystems/scheduler/scheduler.o 00:04:33.292 CC module/event/subsystems/sock/sock.o 00:04:33.552 LIB libspdk_event_vhost_blk.a 00:04:33.552 LIB libspdk_event_fsdev.a 00:04:33.552 LIB libspdk_event_keyring.a 00:04:33.552 LIB libspdk_event_vmd.a 00:04:33.552 LIB libspdk_event_sock.a 00:04:33.552 LIB libspdk_event_scheduler.a 00:04:33.552 LIB libspdk_event_iobuf.a 00:04:33.552 SO libspdk_event_vhost_blk.so.3.0 00:04:33.552 SO libspdk_event_fsdev.so.1.0 00:04:33.552 SO libspdk_event_keyring.so.1.0 00:04:33.552 SO libspdk_event_sock.so.5.0 00:04:33.552 SO libspdk_event_vmd.so.6.0 00:04:33.552 SO libspdk_event_scheduler.so.4.0 00:04:33.552 SO libspdk_event_iobuf.so.3.0 00:04:33.552 SYMLINK libspdk_event_fsdev.so 00:04:33.552 SYMLINK libspdk_event_keyring.so 00:04:33.552 SYMLINK libspdk_event_vhost_blk.so 00:04:33.552 SYMLINK libspdk_event_scheduler.so 00:04:33.552 SYMLINK libspdk_event_sock.so 00:04:33.552 SYMLINK libspdk_event_iobuf.so 00:04:33.552 SYMLINK libspdk_event_vmd.so 00:04:33.811 CC module/event/subsystems/accel/accel.o 00:04:34.071 LIB libspdk_event_accel.a 00:04:34.071 SO libspdk_event_accel.so.6.0 00:04:34.071 SYMLINK libspdk_event_accel.so 00:04:34.640 CC module/event/subsystems/bdev/bdev.o 00:04:34.640 LIB libspdk_event_bdev.a 00:04:34.641 SO libspdk_event_bdev.so.6.0 00:04:34.900 SYMLINK libspdk_event_bdev.so 00:04:35.160 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:35.160 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:35.160 CC module/event/subsystems/scsi/scsi.o 00:04:35.160 CC module/event/subsystems/ublk/ublk.o 00:04:35.160 CC module/event/subsystems/nbd/nbd.o 00:04:35.419 LIB libspdk_event_scsi.a 00:04:35.419 LIB libspdk_event_ublk.a 00:04:35.419 LIB libspdk_event_nbd.a 00:04:35.419 SO libspdk_event_scsi.so.6.0 00:04:35.419 SO libspdk_event_nbd.so.6.0 00:04:35.419 SO libspdk_event_ublk.so.3.0 00:04:35.419 SYMLINK libspdk_event_scsi.so 00:04:35.419 SYMLINK libspdk_event_nbd.so 00:04:35.419 LIB libspdk_event_nvmf.a 00:04:35.419 SYMLINK libspdk_event_ublk.so 00:04:35.419 SO libspdk_event_nvmf.so.6.0 00:04:35.419 SYMLINK libspdk_event_nvmf.so 00:04:35.679 CC module/event/subsystems/iscsi/iscsi.o 00:04:35.679 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:35.939 LIB libspdk_event_vhost_scsi.a 00:04:35.939 LIB libspdk_event_iscsi.a 00:04:35.939 SO libspdk_event_vhost_scsi.so.3.0 00:04:35.939 SO libspdk_event_iscsi.so.6.0 00:04:35.939 SYMLINK libspdk_event_vhost_scsi.so 00:04:35.939 SYMLINK libspdk_event_iscsi.so 00:04:36.199 SO libspdk.so.6.0 00:04:36.199 SYMLINK libspdk.so 00:04:36.458 TEST_HEADER include/spdk/accel.h 00:04:36.458 TEST_HEADER include/spdk/accel_module.h 00:04:36.458 TEST_HEADER include/spdk/assert.h 00:04:36.458 TEST_HEADER include/spdk/barrier.h 00:04:36.458 CXX app/trace/trace.o 00:04:36.458 TEST_HEADER include/spdk/base64.h 00:04:36.458 CC test/rpc_client/rpc_client_test.o 00:04:36.458 TEST_HEADER include/spdk/bdev.h 00:04:36.458 TEST_HEADER include/spdk/bdev_module.h 00:04:36.458 TEST_HEADER include/spdk/bdev_zone.h 00:04:36.458 TEST_HEADER include/spdk/bit_array.h 00:04:36.458 TEST_HEADER include/spdk/bit_pool.h 00:04:36.458 TEST_HEADER include/spdk/blob_bdev.h 00:04:36.459 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:36.459 TEST_HEADER include/spdk/blobfs.h 00:04:36.459 TEST_HEADER include/spdk/blob.h 00:04:36.459 TEST_HEADER include/spdk/conf.h 00:04:36.459 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:36.459 TEST_HEADER include/spdk/config.h 00:04:36.459 TEST_HEADER include/spdk/cpuset.h 00:04:36.459 TEST_HEADER include/spdk/crc16.h 00:04:36.459 TEST_HEADER include/spdk/crc32.h 00:04:36.459 TEST_HEADER include/spdk/crc64.h 00:04:36.459 TEST_HEADER include/spdk/dif.h 00:04:36.459 TEST_HEADER include/spdk/dma.h 00:04:36.459 TEST_HEADER include/spdk/endian.h 00:04:36.459 TEST_HEADER include/spdk/env_dpdk.h 00:04:36.459 TEST_HEADER include/spdk/env.h 00:04:36.459 TEST_HEADER include/spdk/event.h 00:04:36.459 TEST_HEADER include/spdk/fd_group.h 00:04:36.459 TEST_HEADER include/spdk/fd.h 00:04:36.459 CC test/thread/poller_perf/poller_perf.o 00:04:36.459 TEST_HEADER include/spdk/file.h 00:04:36.459 TEST_HEADER include/spdk/fsdev.h 00:04:36.459 TEST_HEADER include/spdk/fsdev_module.h 00:04:36.459 TEST_HEADER include/spdk/ftl.h 00:04:36.459 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:36.459 TEST_HEADER include/spdk/gpt_spec.h 00:04:36.459 TEST_HEADER include/spdk/hexlify.h 00:04:36.459 CC examples/ioat/perf/perf.o 00:04:36.459 TEST_HEADER include/spdk/histogram_data.h 00:04:36.459 CC examples/util/zipf/zipf.o 00:04:36.459 TEST_HEADER include/spdk/idxd.h 00:04:36.459 TEST_HEADER include/spdk/idxd_spec.h 00:04:36.459 TEST_HEADER include/spdk/init.h 00:04:36.459 TEST_HEADER include/spdk/ioat.h 00:04:36.459 TEST_HEADER include/spdk/ioat_spec.h 00:04:36.459 TEST_HEADER include/spdk/iscsi_spec.h 00:04:36.459 TEST_HEADER include/spdk/json.h 00:04:36.459 TEST_HEADER include/spdk/jsonrpc.h 00:04:36.459 TEST_HEADER include/spdk/keyring.h 00:04:36.459 TEST_HEADER include/spdk/keyring_module.h 00:04:36.459 CC test/dma/test_dma/test_dma.o 00:04:36.718 TEST_HEADER include/spdk/likely.h 00:04:36.718 TEST_HEADER include/spdk/log.h 00:04:36.718 TEST_HEADER include/spdk/lvol.h 00:04:36.718 TEST_HEADER include/spdk/md5.h 00:04:36.718 TEST_HEADER include/spdk/memory.h 00:04:36.718 CC test/app/bdev_svc/bdev_svc.o 00:04:36.718 TEST_HEADER include/spdk/mmio.h 00:04:36.718 TEST_HEADER include/spdk/nbd.h 00:04:36.718 TEST_HEADER include/spdk/net.h 00:04:36.718 TEST_HEADER include/spdk/notify.h 00:04:36.718 TEST_HEADER include/spdk/nvme.h 00:04:36.719 TEST_HEADER include/spdk/nvme_intel.h 00:04:36.719 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:36.719 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:36.719 TEST_HEADER include/spdk/nvme_spec.h 00:04:36.719 TEST_HEADER include/spdk/nvme_zns.h 00:04:36.719 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:36.719 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:36.719 TEST_HEADER include/spdk/nvmf.h 00:04:36.719 TEST_HEADER include/spdk/nvmf_spec.h 00:04:36.719 TEST_HEADER include/spdk/nvmf_transport.h 00:04:36.719 TEST_HEADER include/spdk/opal.h 00:04:36.719 TEST_HEADER include/spdk/opal_spec.h 00:04:36.719 CC test/env/mem_callbacks/mem_callbacks.o 00:04:36.719 TEST_HEADER include/spdk/pci_ids.h 00:04:36.719 TEST_HEADER include/spdk/pipe.h 00:04:36.719 TEST_HEADER include/spdk/queue.h 00:04:36.719 TEST_HEADER include/spdk/reduce.h 00:04:36.719 TEST_HEADER include/spdk/rpc.h 00:04:36.719 TEST_HEADER include/spdk/scheduler.h 00:04:36.719 TEST_HEADER include/spdk/scsi.h 00:04:36.719 TEST_HEADER include/spdk/scsi_spec.h 00:04:36.719 TEST_HEADER include/spdk/sock.h 00:04:36.719 LINK rpc_client_test 00:04:36.719 LINK interrupt_tgt 00:04:36.719 TEST_HEADER include/spdk/stdinc.h 00:04:36.719 TEST_HEADER include/spdk/string.h 00:04:36.719 LINK poller_perf 00:04:36.719 TEST_HEADER include/spdk/thread.h 00:04:36.719 TEST_HEADER include/spdk/trace.h 00:04:36.719 TEST_HEADER include/spdk/trace_parser.h 00:04:36.719 TEST_HEADER include/spdk/tree.h 00:04:36.719 TEST_HEADER include/spdk/ublk.h 00:04:36.719 TEST_HEADER include/spdk/util.h 00:04:36.719 TEST_HEADER include/spdk/uuid.h 00:04:36.719 TEST_HEADER include/spdk/version.h 00:04:36.719 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:36.719 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:36.719 TEST_HEADER include/spdk/vhost.h 00:04:36.719 TEST_HEADER include/spdk/vmd.h 00:04:36.719 LINK zipf 00:04:36.719 TEST_HEADER include/spdk/xor.h 00:04:36.719 TEST_HEADER include/spdk/zipf.h 00:04:36.719 CXX test/cpp_headers/accel.o 00:04:36.719 LINK bdev_svc 00:04:36.719 LINK ioat_perf 00:04:36.979 LINK spdk_trace 00:04:36.979 CXX test/cpp_headers/accel_module.o 00:04:36.979 CC app/trace_record/trace_record.o 00:04:36.979 CC app/nvmf_tgt/nvmf_main.o 00:04:36.979 CC app/iscsi_tgt/iscsi_tgt.o 00:04:36.979 CXX test/cpp_headers/assert.o 00:04:36.979 CC app/spdk_tgt/spdk_tgt.o 00:04:36.979 CC examples/ioat/verify/verify.o 00:04:36.979 LINK test_dma 00:04:37.239 CC app/spdk_lspci/spdk_lspci.o 00:04:37.239 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:37.239 CXX test/cpp_headers/barrier.o 00:04:37.239 LINK mem_callbacks 00:04:37.239 LINK nvmf_tgt 00:04:37.239 LINK iscsi_tgt 00:04:37.239 LINK spdk_trace_record 00:04:37.239 LINK spdk_tgt 00:04:37.239 LINK verify 00:04:37.239 LINK spdk_lspci 00:04:37.239 CXX test/cpp_headers/base64.o 00:04:37.499 CC test/env/vtophys/vtophys.o 00:04:37.499 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:37.499 CXX test/cpp_headers/bdev.o 00:04:37.499 CXX test/cpp_headers/bdev_module.o 00:04:37.499 CC test/app/jsoncat/jsoncat.o 00:04:37.499 CC test/app/histogram_perf/histogram_perf.o 00:04:37.499 CC test/app/stub/stub.o 00:04:37.499 LINK vtophys 00:04:37.499 CC app/spdk_nvme_perf/perf.o 00:04:37.499 LINK jsoncat 00:04:37.499 CC examples/thread/thread/thread_ex.o 00:04:37.499 LINK histogram_perf 00:04:37.499 LINK nvme_fuzz 00:04:37.759 CXX test/cpp_headers/bdev_zone.o 00:04:37.759 LINK stub 00:04:37.759 CC app/spdk_nvme_identify/identify.o 00:04:37.759 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:37.759 CC app/spdk_nvme_discover/discovery_aer.o 00:04:37.759 CXX test/cpp_headers/bit_array.o 00:04:37.759 CC app/spdk_top/spdk_top.o 00:04:37.759 LINK thread 00:04:38.019 LINK env_dpdk_post_init 00:04:38.019 CC examples/sock/hello_world/hello_sock.o 00:04:38.019 CC app/vhost/vhost.o 00:04:38.019 CXX test/cpp_headers/bit_pool.o 00:04:38.019 LINK spdk_nvme_discover 00:04:38.019 CXX test/cpp_headers/blob_bdev.o 00:04:38.291 LINK vhost 00:04:38.291 CC test/env/memory/memory_ut.o 00:04:38.291 LINK hello_sock 00:04:38.291 CC app/spdk_dd/spdk_dd.o 00:04:38.291 CXX test/cpp_headers/blobfs_bdev.o 00:04:38.291 CC app/fio/nvme/fio_plugin.o 00:04:38.555 CC test/env/pci/pci_ut.o 00:04:38.555 CXX test/cpp_headers/blobfs.o 00:04:38.555 CC examples/vmd/lsvmd/lsvmd.o 00:04:38.555 LINK spdk_nvme_perf 00:04:38.555 LINK spdk_dd 00:04:38.555 LINK spdk_nvme_identify 00:04:38.555 LINK lsvmd 00:04:38.555 CXX test/cpp_headers/blob.o 00:04:38.815 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:38.815 LINK spdk_top 00:04:38.815 CXX test/cpp_headers/conf.o 00:04:38.815 CC examples/vmd/led/led.o 00:04:38.815 LINK spdk_nvme 00:04:38.815 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:38.815 LINK pci_ut 00:04:39.075 CXX test/cpp_headers/config.o 00:04:39.075 CC test/event/event_perf/event_perf.o 00:04:39.075 CC test/nvme/aer/aer.o 00:04:39.075 CXX test/cpp_headers/cpuset.o 00:04:39.075 CC test/nvme/reset/reset.o 00:04:39.075 LINK led 00:04:39.075 CC app/fio/bdev/fio_plugin.o 00:04:39.075 LINK event_perf 00:04:39.075 CXX test/cpp_headers/crc16.o 00:04:39.334 LINK iscsi_fuzz 00:04:39.334 LINK memory_ut 00:04:39.334 LINK reset 00:04:39.334 LINK aer 00:04:39.334 LINK vhost_fuzz 00:04:39.334 CXX test/cpp_headers/crc32.o 00:04:39.334 CC examples/idxd/perf/perf.o 00:04:39.335 CC test/event/reactor/reactor.o 00:04:39.335 CC test/accel/dif/dif.o 00:04:39.595 CXX test/cpp_headers/crc64.o 00:04:39.595 CXX test/cpp_headers/dif.o 00:04:39.595 LINK reactor 00:04:39.595 CC test/nvme/e2edp/nvme_dp.o 00:04:39.595 CC test/nvme/sgl/sgl.o 00:04:39.595 CC test/blobfs/mkfs/mkfs.o 00:04:39.595 LINK spdk_bdev 00:04:39.595 CXX test/cpp_headers/dma.o 00:04:39.595 CC test/lvol/esnap/esnap.o 00:04:39.595 LINK idxd_perf 00:04:39.855 CC test/nvme/overhead/overhead.o 00:04:39.855 CC test/event/reactor_perf/reactor_perf.o 00:04:39.855 LINK mkfs 00:04:39.855 CXX test/cpp_headers/endian.o 00:04:39.855 LINK sgl 00:04:39.855 LINK nvme_dp 00:04:39.855 CC test/event/app_repeat/app_repeat.o 00:04:39.855 LINK reactor_perf 00:04:39.855 CXX test/cpp_headers/env_dpdk.o 00:04:40.115 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:40.115 LINK app_repeat 00:04:40.115 LINK overhead 00:04:40.115 CC test/nvme/err_injection/err_injection.o 00:04:40.115 CXX test/cpp_headers/env.o 00:04:40.115 CC test/event/scheduler/scheduler.o 00:04:40.115 CC examples/accel/perf/accel_perf.o 00:04:40.115 LINK dif 00:04:40.115 CC examples/blob/hello_world/hello_blob.o 00:04:40.115 CXX test/cpp_headers/event.o 00:04:40.374 CC test/nvme/startup/startup.o 00:04:40.374 LINK hello_fsdev 00:04:40.374 CC test/nvme/reserve/reserve.o 00:04:40.374 LINK err_injection 00:04:40.374 CXX test/cpp_headers/fd_group.o 00:04:40.374 LINK scheduler 00:04:40.374 LINK hello_blob 00:04:40.374 LINK startup 00:04:40.374 CC test/nvme/simple_copy/simple_copy.o 00:04:40.374 CXX test/cpp_headers/fd.o 00:04:40.374 CXX test/cpp_headers/file.o 00:04:40.634 LINK reserve 00:04:40.634 CXX test/cpp_headers/fsdev.o 00:04:40.634 CC test/bdev/bdevio/bdevio.o 00:04:40.634 CC test/nvme/connect_stress/connect_stress.o 00:04:40.634 LINK simple_copy 00:04:40.634 CC test/nvme/boot_partition/boot_partition.o 00:04:40.634 LINK accel_perf 00:04:40.634 CC test/nvme/compliance/nvme_compliance.o 00:04:40.634 CC examples/blob/cli/blobcli.o 00:04:40.634 CC test/nvme/fused_ordering/fused_ordering.o 00:04:40.894 CXX test/cpp_headers/fsdev_module.o 00:04:40.894 LINK boot_partition 00:04:40.894 LINK connect_stress 00:04:40.894 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:40.894 CC test/nvme/fdp/fdp.o 00:04:40.894 CXX test/cpp_headers/ftl.o 00:04:40.894 LINK fused_ordering 00:04:41.153 LINK bdevio 00:04:41.153 LINK nvme_compliance 00:04:41.153 CC test/nvme/cuse/cuse.o 00:04:41.153 LINK doorbell_aers 00:04:41.153 CXX test/cpp_headers/fuse_dispatcher.o 00:04:41.153 CC examples/nvme/hello_world/hello_world.o 00:04:41.153 CC examples/nvme/reconnect/reconnect.o 00:04:41.153 CXX test/cpp_headers/gpt_spec.o 00:04:41.153 LINK blobcli 00:04:41.153 CXX test/cpp_headers/hexlify.o 00:04:41.153 CXX test/cpp_headers/histogram_data.o 00:04:41.153 CXX test/cpp_headers/idxd.o 00:04:41.153 LINK fdp 00:04:41.413 CXX test/cpp_headers/idxd_spec.o 00:04:41.413 CXX test/cpp_headers/init.o 00:04:41.413 CXX test/cpp_headers/ioat.o 00:04:41.413 CXX test/cpp_headers/ioat_spec.o 00:04:41.413 LINK hello_world 00:04:41.413 CXX test/cpp_headers/iscsi_spec.o 00:04:41.413 CC examples/bdev/hello_world/hello_bdev.o 00:04:41.413 CXX test/cpp_headers/json.o 00:04:41.413 LINK reconnect 00:04:41.413 CXX test/cpp_headers/jsonrpc.o 00:04:41.413 CXX test/cpp_headers/keyring.o 00:04:41.673 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:41.673 CC examples/bdev/bdevperf/bdevperf.o 00:04:41.673 CC examples/nvme/arbitration/arbitration.o 00:04:41.673 CXX test/cpp_headers/keyring_module.o 00:04:41.673 CXX test/cpp_headers/likely.o 00:04:41.673 LINK hello_bdev 00:04:41.673 CC examples/nvme/hotplug/hotplug.o 00:04:41.673 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:41.933 CXX test/cpp_headers/log.o 00:04:41.933 CC examples/nvme/abort/abort.o 00:04:41.933 LINK arbitration 00:04:41.933 CXX test/cpp_headers/lvol.o 00:04:41.933 CXX test/cpp_headers/md5.o 00:04:41.933 LINK cmb_copy 00:04:41.933 LINK hotplug 00:04:42.192 LINK nvme_manage 00:04:42.192 CXX test/cpp_headers/memory.o 00:04:42.192 CXX test/cpp_headers/mmio.o 00:04:42.192 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:42.192 CXX test/cpp_headers/nbd.o 00:04:42.192 CXX test/cpp_headers/net.o 00:04:42.192 CXX test/cpp_headers/notify.o 00:04:42.192 LINK abort 00:04:42.192 CXX test/cpp_headers/nvme.o 00:04:42.192 CXX test/cpp_headers/nvme_intel.o 00:04:42.192 CXX test/cpp_headers/nvme_ocssd.o 00:04:42.192 LINK cuse 00:04:42.192 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:42.452 LINK pmr_persistence 00:04:42.452 CXX test/cpp_headers/nvme_spec.o 00:04:42.452 CXX test/cpp_headers/nvme_zns.o 00:04:42.452 CXX test/cpp_headers/nvmf_cmd.o 00:04:42.452 LINK bdevperf 00:04:42.452 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:42.452 CXX test/cpp_headers/nvmf.o 00:04:42.452 CXX test/cpp_headers/nvmf_spec.o 00:04:42.452 CXX test/cpp_headers/nvmf_transport.o 00:04:42.452 CXX test/cpp_headers/opal.o 00:04:42.452 CXX test/cpp_headers/opal_spec.o 00:04:42.452 CXX test/cpp_headers/pci_ids.o 00:04:42.712 CXX test/cpp_headers/pipe.o 00:04:42.712 CXX test/cpp_headers/queue.o 00:04:42.712 CXX test/cpp_headers/reduce.o 00:04:42.712 CXX test/cpp_headers/rpc.o 00:04:42.712 CXX test/cpp_headers/scheduler.o 00:04:42.712 CXX test/cpp_headers/scsi.o 00:04:42.712 CXX test/cpp_headers/scsi_spec.o 00:04:42.712 CXX test/cpp_headers/sock.o 00:04:42.712 CXX test/cpp_headers/stdinc.o 00:04:42.712 CXX test/cpp_headers/string.o 00:04:42.712 CXX test/cpp_headers/thread.o 00:04:42.712 CXX test/cpp_headers/trace.o 00:04:42.712 CXX test/cpp_headers/trace_parser.o 00:04:42.712 CXX test/cpp_headers/tree.o 00:04:42.971 CC examples/nvmf/nvmf/nvmf.o 00:04:42.971 CXX test/cpp_headers/ublk.o 00:04:42.971 CXX test/cpp_headers/util.o 00:04:42.971 CXX test/cpp_headers/uuid.o 00:04:42.971 CXX test/cpp_headers/version.o 00:04:42.971 CXX test/cpp_headers/vfio_user_pci.o 00:04:42.971 CXX test/cpp_headers/vfio_user_spec.o 00:04:42.971 CXX test/cpp_headers/vhost.o 00:04:42.971 CXX test/cpp_headers/vmd.o 00:04:42.971 CXX test/cpp_headers/xor.o 00:04:42.971 CXX test/cpp_headers/zipf.o 00:04:43.230 LINK nvmf 00:04:45.781 LINK esnap 00:04:45.781 ************************************ 00:04:45.781 END TEST make 00:04:45.781 ************************************ 00:04:45.781 00:04:45.781 real 1m14.202s 00:04:45.781 user 5m41.939s 00:04:45.781 sys 1m8.886s 00:04:45.781 14:03:50 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:04:45.781 14:03:50 make -- common/autotest_common.sh@10 -- $ set +x 00:04:45.781 14:03:50 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:04:45.781 14:03:50 -- pm/common@29 -- $ signal_monitor_resources TERM 00:04:45.781 14:03:50 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:04:45.781 14:03:50 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.781 14:03:50 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:04:45.781 14:03:50 -- pm/common@44 -- $ pid=6202 00:04:45.781 14:03:50 -- pm/common@50 -- $ kill -TERM 6202 00:04:45.781 14:03:50 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.781 14:03:50 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:04:45.781 14:03:50 -- pm/common@44 -- $ pid=6204 00:04:45.781 14:03:50 -- pm/common@50 -- $ kill -TERM 6204 00:04:46.040 14:03:50 -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:04:46.040 14:03:50 -- common/autotest_common.sh@1681 -- # lcov --version 00:04:46.040 14:03:50 -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:04:46.040 14:03:50 -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:04:46.040 14:03:50 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:46.040 14:03:50 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:46.040 14:03:50 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:46.040 14:03:50 -- scripts/common.sh@336 -- # IFS=.-: 00:04:46.040 14:03:50 -- scripts/common.sh@336 -- # read -ra ver1 00:04:46.040 14:03:50 -- scripts/common.sh@337 -- # IFS=.-: 00:04:46.040 14:03:50 -- scripts/common.sh@337 -- # read -ra ver2 00:04:46.040 14:03:50 -- scripts/common.sh@338 -- # local 'op=<' 00:04:46.040 14:03:50 -- scripts/common.sh@340 -- # ver1_l=2 00:04:46.040 14:03:50 -- scripts/common.sh@341 -- # ver2_l=1 00:04:46.040 14:03:50 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:46.040 14:03:50 -- scripts/common.sh@344 -- # case "$op" in 00:04:46.040 14:03:50 -- scripts/common.sh@345 -- # : 1 00:04:46.040 14:03:50 -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:46.040 14:03:50 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:46.040 14:03:50 -- scripts/common.sh@365 -- # decimal 1 00:04:46.040 14:03:50 -- scripts/common.sh@353 -- # local d=1 00:04:46.040 14:03:50 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:46.040 14:03:50 -- scripts/common.sh@355 -- # echo 1 00:04:46.040 14:03:50 -- scripts/common.sh@365 -- # ver1[v]=1 00:04:46.040 14:03:50 -- scripts/common.sh@366 -- # decimal 2 00:04:46.040 14:03:50 -- scripts/common.sh@353 -- # local d=2 00:04:46.040 14:03:50 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:46.040 14:03:50 -- scripts/common.sh@355 -- # echo 2 00:04:46.040 14:03:50 -- scripts/common.sh@366 -- # ver2[v]=2 00:04:46.040 14:03:50 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:46.040 14:03:50 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:46.040 14:03:50 -- scripts/common.sh@368 -- # return 0 00:04:46.040 14:03:50 -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:46.040 14:03:50 -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:04:46.040 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:46.040 --rc genhtml_branch_coverage=1 00:04:46.040 --rc genhtml_function_coverage=1 00:04:46.040 --rc genhtml_legend=1 00:04:46.040 --rc geninfo_all_blocks=1 00:04:46.040 --rc geninfo_unexecuted_blocks=1 00:04:46.040 00:04:46.040 ' 00:04:46.040 14:03:50 -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:04:46.040 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:46.040 --rc genhtml_branch_coverage=1 00:04:46.040 --rc genhtml_function_coverage=1 00:04:46.040 --rc genhtml_legend=1 00:04:46.040 --rc geninfo_all_blocks=1 00:04:46.040 --rc geninfo_unexecuted_blocks=1 00:04:46.040 00:04:46.040 ' 00:04:46.040 14:03:50 -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:04:46.040 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:46.040 --rc genhtml_branch_coverage=1 00:04:46.040 --rc genhtml_function_coverage=1 00:04:46.040 --rc genhtml_legend=1 00:04:46.040 --rc geninfo_all_blocks=1 00:04:46.040 --rc geninfo_unexecuted_blocks=1 00:04:46.040 00:04:46.040 ' 00:04:46.040 14:03:50 -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:04:46.040 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:46.040 --rc genhtml_branch_coverage=1 00:04:46.040 --rc genhtml_function_coverage=1 00:04:46.040 --rc genhtml_legend=1 00:04:46.040 --rc geninfo_all_blocks=1 00:04:46.040 --rc geninfo_unexecuted_blocks=1 00:04:46.040 00:04:46.040 ' 00:04:46.040 14:03:50 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:04:46.040 14:03:50 -- nvmf/common.sh@7 -- # uname -s 00:04:46.040 14:03:50 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:46.040 14:03:50 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:46.040 14:03:50 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:46.040 14:03:50 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:46.041 14:03:50 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:46.041 14:03:50 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:46.041 14:03:50 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:46.041 14:03:50 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:46.041 14:03:50 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:46.041 14:03:50 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:46.041 14:03:50 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:a1e24555-614e-41f5-8ac1-aaa73d42684b 00:04:46.041 14:03:50 -- nvmf/common.sh@18 -- # NVME_HOSTID=a1e24555-614e-41f5-8ac1-aaa73d42684b 00:04:46.041 14:03:50 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:46.041 14:03:50 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:46.041 14:03:50 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:46.041 14:03:50 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:46.041 14:03:50 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:04:46.041 14:03:50 -- scripts/common.sh@15 -- # shopt -s extglob 00:04:46.041 14:03:50 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:46.041 14:03:50 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:46.041 14:03:50 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:46.041 14:03:50 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:46.041 14:03:50 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:46.041 14:03:50 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:46.041 14:03:50 -- paths/export.sh@5 -- # export PATH 00:04:46.041 14:03:50 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:46.041 14:03:50 -- nvmf/common.sh@51 -- # : 0 00:04:46.041 14:03:50 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:46.041 14:03:50 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:46.041 14:03:50 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:46.041 14:03:50 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:46.041 14:03:50 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:46.041 14:03:50 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:46.041 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:46.041 14:03:50 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:46.041 14:03:50 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:46.041 14:03:50 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:46.041 14:03:50 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:04:46.041 14:03:50 -- spdk/autotest.sh@32 -- # uname -s 00:04:46.041 14:03:50 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:04:46.041 14:03:50 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:04:46.041 14:03:50 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:46.041 14:03:50 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:04:46.041 14:03:50 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:46.041 14:03:50 -- spdk/autotest.sh@44 -- # modprobe nbd 00:04:46.041 14:03:50 -- spdk/autotest.sh@46 -- # type -P udevadm 00:04:46.041 14:03:50 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:04:46.041 14:03:50 -- spdk/autotest.sh@48 -- # udevadm_pid=67799 00:04:46.041 14:03:50 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:04:46.041 14:03:50 -- pm/common@17 -- # local monitor 00:04:46.041 14:03:50 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:46.041 14:03:50 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:46.041 14:03:50 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:04:46.041 14:03:50 -- pm/common@25 -- # sleep 1 00:04:46.041 14:03:50 -- pm/common@21 -- # date +%s 00:04:46.041 14:03:50 -- pm/common@21 -- # date +%s 00:04:46.041 14:03:50 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727705030 00:04:46.041 14:03:50 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727705030 00:04:46.301 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727705030_collect-vmstat.pm.log 00:04:46.301 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727705030_collect-cpu-load.pm.log 00:04:47.242 14:03:51 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:04:47.242 14:03:51 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:04:47.242 14:03:51 -- common/autotest_common.sh@724 -- # xtrace_disable 00:04:47.242 14:03:51 -- common/autotest_common.sh@10 -- # set +x 00:04:47.242 14:03:51 -- spdk/autotest.sh@59 -- # create_test_list 00:04:47.242 14:03:51 -- common/autotest_common.sh@748 -- # xtrace_disable 00:04:47.242 14:03:51 -- common/autotest_common.sh@10 -- # set +x 00:04:47.242 14:03:51 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:04:47.242 14:03:51 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:04:47.242 14:03:51 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:04:47.242 14:03:51 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:04:47.242 14:03:51 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:04:47.242 14:03:51 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:04:47.242 14:03:51 -- common/autotest_common.sh@1455 -- # uname 00:04:47.242 14:03:51 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:04:47.242 14:03:51 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:04:47.242 14:03:51 -- common/autotest_common.sh@1475 -- # uname 00:04:47.242 14:03:51 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:04:47.242 14:03:51 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:04:47.242 14:03:51 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:04:47.242 lcov: LCOV version 1.15 00:04:47.242 14:03:51 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:05:02.135 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:02.135 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:17.028 14:04:19 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:17.028 14:04:19 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:17.028 14:04:19 -- common/autotest_common.sh@10 -- # set +x 00:05:17.028 14:04:19 -- spdk/autotest.sh@78 -- # rm -f 00:05:17.028 14:04:19 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:17.028 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:17.028 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:17.028 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:17.028 14:04:20 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:17.028 14:04:20 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:05:17.028 14:04:20 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:05:17.028 14:04:20 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:05:17.028 14:04:20 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:17.028 14:04:20 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:05:17.028 14:04:20 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:05:17.028 14:04:20 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:17.028 14:04:20 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n1 00:05:17.028 14:04:20 -- common/autotest_common.sh@1648 -- # local device=nvme1n1 00:05:17.028 14:04:20 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:17.028 14:04:20 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n2 00:05:17.028 14:04:20 -- common/autotest_common.sh@1648 -- # local device=nvme1n2 00:05:17.028 14:04:20 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:17.028 14:04:20 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n3 00:05:17.028 14:04:20 -- common/autotest_common.sh@1648 -- # local device=nvme1n3 00:05:17.028 14:04:20 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:05:17.028 14:04:20 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:17.028 14:04:20 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:17.028 14:04:20 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:17.028 14:04:20 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:17.028 14:04:20 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:17.028 14:04:20 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:17.028 14:04:20 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:17.028 No valid GPT data, bailing 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # pt= 00:05:17.028 14:04:20 -- scripts/common.sh@395 -- # return 1 00:05:17.028 14:04:20 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:17.028 1+0 records in 00:05:17.028 1+0 records out 00:05:17.028 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00668729 s, 157 MB/s 00:05:17.028 14:04:20 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:17.028 14:04:20 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:17.028 14:04:20 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:17.028 14:04:20 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:17.028 14:04:20 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:17.028 No valid GPT data, bailing 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # pt= 00:05:17.028 14:04:20 -- scripts/common.sh@395 -- # return 1 00:05:17.028 14:04:20 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:17.028 1+0 records in 00:05:17.028 1+0 records out 00:05:17.028 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00656798 s, 160 MB/s 00:05:17.028 14:04:20 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:17.028 14:04:20 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:17.028 14:04:20 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:05:17.028 14:04:20 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:05:17.028 14:04:20 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:05:17.028 No valid GPT data, bailing 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:05:17.028 14:04:20 -- scripts/common.sh@394 -- # pt= 00:05:17.028 14:04:20 -- scripts/common.sh@395 -- # return 1 00:05:17.028 14:04:20 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:05:17.028 1+0 records in 00:05:17.028 1+0 records out 00:05:17.028 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00648936 s, 162 MB/s 00:05:17.028 14:04:20 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:17.028 14:04:20 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:17.028 14:04:20 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:05:17.028 14:04:20 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:05:17.028 14:04:20 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:05:17.028 No valid GPT data, bailing 00:05:17.028 14:04:21 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:05:17.028 14:04:21 -- scripts/common.sh@394 -- # pt= 00:05:17.028 14:04:21 -- scripts/common.sh@395 -- # return 1 00:05:17.028 14:04:21 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:05:17.028 1+0 records in 00:05:17.028 1+0 records out 00:05:17.028 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00698848 s, 150 MB/s 00:05:17.028 14:04:21 -- spdk/autotest.sh@105 -- # sync 00:05:17.028 14:04:21 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:17.028 14:04:21 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:17.028 14:04:21 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:19.567 14:04:24 -- spdk/autotest.sh@111 -- # uname -s 00:05:19.567 14:04:24 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:19.567 14:04:24 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:19.567 14:04:24 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:20.505 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:20.505 Hugepages 00:05:20.505 node hugesize free / total 00:05:20.505 node0 1048576kB 0 / 0 00:05:20.505 node0 2048kB 0 / 0 00:05:20.505 00:05:20.505 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:20.505 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:20.505 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:05:20.765 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:05:20.765 14:04:25 -- spdk/autotest.sh@117 -- # uname -s 00:05:20.765 14:04:25 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:20.765 14:04:25 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:20.765 14:04:25 -- common/autotest_common.sh@1514 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:21.333 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:21.593 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:21.593 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:21.593 14:04:26 -- common/autotest_common.sh@1515 -- # sleep 1 00:05:22.974 14:04:27 -- common/autotest_common.sh@1516 -- # bdfs=() 00:05:22.974 14:04:27 -- common/autotest_common.sh@1516 -- # local bdfs 00:05:22.974 14:04:27 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:05:22.974 14:04:27 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:05:22.974 14:04:27 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:22.974 14:04:27 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:22.974 14:04:27 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:22.974 14:04:27 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:22.974 14:04:27 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:22.974 14:04:27 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:22.974 14:04:27 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:22.974 14:04:27 -- common/autotest_common.sh@1520 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:23.234 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:23.234 Waiting for block devices as requested 00:05:23.234 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:23.494 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:23.494 14:04:28 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:23.494 14:04:28 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # grep 0000:00:10.0/nvme/nvme 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme1 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:23.494 14:04:28 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:23.494 14:04:28 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:23.494 14:04:28 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1541 -- # continue 00:05:23.494 14:04:28 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:23.494 14:04:28 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # grep 0000:00:11.0/nvme/nvme 00:05:23.494 14:04:28 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:23.494 14:04:28 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:23.494 14:04:28 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:23.494 14:04:28 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:23.494 14:04:28 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:23.494 14:04:28 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:23.494 14:04:28 -- common/autotest_common.sh@1541 -- # continue 00:05:23.494 14:04:28 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:23.494 14:04:28 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:23.494 14:04:28 -- common/autotest_common.sh@10 -- # set +x 00:05:23.755 14:04:28 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:23.755 14:04:28 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:23.755 14:04:28 -- common/autotest_common.sh@10 -- # set +x 00:05:23.755 14:04:28 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:24.325 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:24.584 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:24.584 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:24.584 14:04:29 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:24.584 14:04:29 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:24.584 14:04:29 -- common/autotest_common.sh@10 -- # set +x 00:05:24.584 14:04:29 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:24.584 14:04:29 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:05:24.584 14:04:29 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:05:24.584 14:04:29 -- common/autotest_common.sh@1561 -- # bdfs=() 00:05:24.584 14:04:29 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:05:24.584 14:04:29 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:05:24.584 14:04:29 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:05:24.584 14:04:29 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:05:24.584 14:04:29 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:24.584 14:04:29 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:24.584 14:04:29 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:24.584 14:04:29 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:24.584 14:04:29 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:24.844 14:04:29 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:24.844 14:04:29 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:24.844 14:04:29 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:24.844 14:04:29 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:24.844 14:04:29 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:24.844 14:04:29 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:24.844 14:04:29 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:24.844 14:04:29 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:24.844 14:04:29 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:24.844 14:04:29 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:24.844 14:04:29 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:05:24.844 14:04:29 -- common/autotest_common.sh@1570 -- # return 0 00:05:24.844 14:04:29 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:05:24.844 14:04:29 -- common/autotest_common.sh@1578 -- # return 0 00:05:24.844 14:04:29 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:24.844 14:04:29 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:24.844 14:04:29 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:24.844 14:04:29 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:24.844 14:04:29 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:24.844 14:04:29 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:24.844 14:04:29 -- common/autotest_common.sh@10 -- # set +x 00:05:24.844 14:04:29 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:24.844 14:04:29 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:24.844 14:04:29 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:24.844 14:04:29 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:24.844 14:04:29 -- common/autotest_common.sh@10 -- # set +x 00:05:24.844 ************************************ 00:05:24.844 START TEST env 00:05:24.844 ************************************ 00:05:24.844 14:04:29 env -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:24.844 * Looking for test storage... 00:05:24.844 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:24.844 14:04:29 env -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:24.844 14:04:29 env -- common/autotest_common.sh@1681 -- # lcov --version 00:05:24.844 14:04:29 env -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:25.103 14:04:29 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:25.103 14:04:29 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:25.103 14:04:29 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:25.103 14:04:29 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:25.103 14:04:29 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:25.103 14:04:29 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:25.103 14:04:29 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:25.103 14:04:29 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:25.103 14:04:29 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:25.103 14:04:29 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:25.103 14:04:29 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:25.103 14:04:29 env -- scripts/common.sh@344 -- # case "$op" in 00:05:25.103 14:04:29 env -- scripts/common.sh@345 -- # : 1 00:05:25.103 14:04:29 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:25.103 14:04:29 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:25.103 14:04:29 env -- scripts/common.sh@365 -- # decimal 1 00:05:25.103 14:04:29 env -- scripts/common.sh@353 -- # local d=1 00:05:25.103 14:04:29 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:25.103 14:04:29 env -- scripts/common.sh@355 -- # echo 1 00:05:25.103 14:04:29 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:25.103 14:04:29 env -- scripts/common.sh@366 -- # decimal 2 00:05:25.103 14:04:29 env -- scripts/common.sh@353 -- # local d=2 00:05:25.103 14:04:29 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:25.103 14:04:29 env -- scripts/common.sh@355 -- # echo 2 00:05:25.103 14:04:29 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:25.103 14:04:29 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:25.103 14:04:29 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:25.103 14:04:29 env -- scripts/common.sh@368 -- # return 0 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:25.103 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:25.103 --rc genhtml_branch_coverage=1 00:05:25.103 --rc genhtml_function_coverage=1 00:05:25.103 --rc genhtml_legend=1 00:05:25.103 --rc geninfo_all_blocks=1 00:05:25.103 --rc geninfo_unexecuted_blocks=1 00:05:25.103 00:05:25.103 ' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:25.103 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:25.103 --rc genhtml_branch_coverage=1 00:05:25.103 --rc genhtml_function_coverage=1 00:05:25.103 --rc genhtml_legend=1 00:05:25.103 --rc geninfo_all_blocks=1 00:05:25.103 --rc geninfo_unexecuted_blocks=1 00:05:25.103 00:05:25.103 ' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:25.103 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:25.103 --rc genhtml_branch_coverage=1 00:05:25.103 --rc genhtml_function_coverage=1 00:05:25.103 --rc genhtml_legend=1 00:05:25.103 --rc geninfo_all_blocks=1 00:05:25.103 --rc geninfo_unexecuted_blocks=1 00:05:25.103 00:05:25.103 ' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:25.103 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:25.103 --rc genhtml_branch_coverage=1 00:05:25.103 --rc genhtml_function_coverage=1 00:05:25.103 --rc genhtml_legend=1 00:05:25.103 --rc geninfo_all_blocks=1 00:05:25.103 --rc geninfo_unexecuted_blocks=1 00:05:25.103 00:05:25.103 ' 00:05:25.103 14:04:29 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:25.103 14:04:29 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:25.103 14:04:29 env -- common/autotest_common.sh@10 -- # set +x 00:05:25.103 ************************************ 00:05:25.103 START TEST env_memory 00:05:25.103 ************************************ 00:05:25.103 14:04:29 env.env_memory -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:25.103 00:05:25.103 00:05:25.103 CUnit - A unit testing framework for C - Version 2.1-3 00:05:25.103 http://cunit.sourceforge.net/ 00:05:25.103 00:05:25.103 00:05:25.103 Suite: memory 00:05:25.103 Test: alloc and free memory map ...[2024-09-30 14:04:29.641397] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:25.103 passed 00:05:25.103 Test: mem map translation ...[2024-09-30 14:04:29.682523] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:25.103 [2024-09-30 14:04:29.682557] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:25.103 [2024-09-30 14:04:29.682610] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:25.103 [2024-09-30 14:04:29.682628] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:25.103 passed 00:05:25.103 Test: mem map registration ...[2024-09-30 14:04:29.744943] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:25.103 [2024-09-30 14:04:29.745077] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:25.363 passed 00:05:25.363 Test: mem map adjacent registrations ...passed 00:05:25.363 00:05:25.363 Run Summary: Type Total Ran Passed Failed Inactive 00:05:25.363 suites 1 1 n/a 0 0 00:05:25.364 tests 4 4 4 0 0 00:05:25.364 asserts 152 152 152 0 n/a 00:05:25.364 00:05:25.364 Elapsed time = 0.223 seconds 00:05:25.364 00:05:25.364 real 0m0.275s 00:05:25.364 user 0m0.232s 00:05:25.364 sys 0m0.033s 00:05:25.364 14:04:29 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:25.364 14:04:29 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:25.364 ************************************ 00:05:25.364 END TEST env_memory 00:05:25.364 ************************************ 00:05:25.364 14:04:29 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:25.364 14:04:29 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:25.364 14:04:29 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:25.364 14:04:29 env -- common/autotest_common.sh@10 -- # set +x 00:05:25.364 ************************************ 00:05:25.364 START TEST env_vtophys 00:05:25.364 ************************************ 00:05:25.364 14:04:29 env.env_vtophys -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:25.364 EAL: lib.eal log level changed from notice to debug 00:05:25.364 EAL: Detected lcore 0 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 1 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 2 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 3 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 4 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 5 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 6 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 7 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 8 as core 0 on socket 0 00:05:25.364 EAL: Detected lcore 9 as core 0 on socket 0 00:05:25.364 EAL: Maximum logical cores by configuration: 128 00:05:25.364 EAL: Detected CPU lcores: 10 00:05:25.364 EAL: Detected NUMA nodes: 1 00:05:25.364 EAL: Checking presence of .so 'librte_eal.so.25.0' 00:05:25.364 EAL: Detected shared linkage of DPDK 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_pci.so.25.0 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_vdev.so.25.0 00:05:25.364 EAL: Registered [vdev] bus. 00:05:25.364 EAL: bus.vdev log level changed from disabled to notice 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_mempool_ring.so.25.0 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_net_i40e.so.25.0 00:05:25.364 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:25.364 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_pci.so 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_bus_vdev.so 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_mempool_ring.so 00:05:25.364 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-25.0/librte_net_i40e.so 00:05:25.364 EAL: No shared files mode enabled, IPC will be disabled 00:05:25.364 EAL: No shared files mode enabled, IPC is disabled 00:05:25.364 EAL: Selected IOVA mode 'PA' 00:05:25.364 EAL: Probing VFIO support... 00:05:25.364 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:25.364 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:25.364 EAL: Ask a virtual area of 0x2e000 bytes 00:05:25.364 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:25.364 EAL: Setting up physically contiguous memory... 00:05:25.364 EAL: Setting maximum number of open files to 524288 00:05:25.364 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:25.364 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:25.364 EAL: Ask a virtual area of 0x61000 bytes 00:05:25.364 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:25.364 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:25.364 EAL: Ask a virtual area of 0x400000000 bytes 00:05:25.364 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:25.364 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:25.364 EAL: Ask a virtual area of 0x61000 bytes 00:05:25.364 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:25.364 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:25.364 EAL: Ask a virtual area of 0x400000000 bytes 00:05:25.364 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:25.364 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:25.364 EAL: Ask a virtual area of 0x61000 bytes 00:05:25.364 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:25.364 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:25.364 EAL: Ask a virtual area of 0x400000000 bytes 00:05:25.364 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:25.364 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:25.364 EAL: Ask a virtual area of 0x61000 bytes 00:05:25.364 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:25.364 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:25.364 EAL: Ask a virtual area of 0x400000000 bytes 00:05:25.364 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:25.364 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:25.364 EAL: Hugepages will be freed exactly as allocated. 00:05:25.364 EAL: No shared files mode enabled, IPC is disabled 00:05:25.364 EAL: No shared files mode enabled, IPC is disabled 00:05:25.627 EAL: TSC frequency is ~2290000 KHz 00:05:25.627 EAL: Main lcore 0 is ready (tid=7f016f8dba40;cpuset=[0]) 00:05:25.627 EAL: Trying to obtain current memory policy. 00:05:25.627 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.627 EAL: Restoring previous memory policy: 0 00:05:25.627 EAL: request: mp_malloc_sync 00:05:25.627 EAL: No shared files mode enabled, IPC is disabled 00:05:25.627 EAL: Heap on socket 0 was expanded by 2MB 00:05:25.627 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:25.627 EAL: No shared files mode enabled, IPC is disabled 00:05:25.627 EAL: Mem event callback 'spdk:(nil)' registered 00:05:25.627 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:25.627 00:05:25.627 00:05:25.627 CUnit - A unit testing framework for C - Version 2.1-3 00:05:25.627 http://cunit.sourceforge.net/ 00:05:25.627 00:05:25.627 00:05:25.627 Suite: components_suite 00:05:25.888 Test: vtophys_malloc_test ...passed 00:05:25.888 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:25.888 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.888 EAL: Restoring previous memory policy: 4 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was expanded by 4MB 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was shrunk by 4MB 00:05:25.888 EAL: Trying to obtain current memory policy. 00:05:25.888 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.888 EAL: Restoring previous memory policy: 4 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was expanded by 6MB 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was shrunk by 6MB 00:05:25.888 EAL: Trying to obtain current memory policy. 00:05:25.888 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.888 EAL: Restoring previous memory policy: 4 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was expanded by 10MB 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was shrunk by 10MB 00:05:25.888 EAL: Trying to obtain current memory policy. 00:05:25.888 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.888 EAL: Restoring previous memory policy: 4 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.888 EAL: request: mp_malloc_sync 00:05:25.888 EAL: No shared files mode enabled, IPC is disabled 00:05:25.888 EAL: Heap on socket 0 was expanded by 18MB 00:05:25.888 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was shrunk by 18MB 00:05:25.889 EAL: Trying to obtain current memory policy. 00:05:25.889 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.889 EAL: Restoring previous memory policy: 4 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was expanded by 34MB 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was shrunk by 34MB 00:05:25.889 EAL: Trying to obtain current memory policy. 00:05:25.889 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.889 EAL: Restoring previous memory policy: 4 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was expanded by 66MB 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was shrunk by 66MB 00:05:25.889 EAL: Trying to obtain current memory policy. 00:05:25.889 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:25.889 EAL: Restoring previous memory policy: 4 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:25.889 EAL: request: mp_malloc_sync 00:05:25.889 EAL: No shared files mode enabled, IPC is disabled 00:05:25.889 EAL: Heap on socket 0 was expanded by 130MB 00:05:25.889 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.149 EAL: request: mp_malloc_sync 00:05:26.149 EAL: No shared files mode enabled, IPC is disabled 00:05:26.149 EAL: Heap on socket 0 was shrunk by 130MB 00:05:26.149 EAL: Trying to obtain current memory policy. 00:05:26.149 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:26.149 EAL: Restoring previous memory policy: 4 00:05:26.149 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.149 EAL: request: mp_malloc_sync 00:05:26.149 EAL: No shared files mode enabled, IPC is disabled 00:05:26.149 EAL: Heap on socket 0 was expanded by 258MB 00:05:26.149 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.149 EAL: request: mp_malloc_sync 00:05:26.149 EAL: No shared files mode enabled, IPC is disabled 00:05:26.149 EAL: Heap on socket 0 was shrunk by 258MB 00:05:26.149 EAL: Trying to obtain current memory policy. 00:05:26.149 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:26.149 EAL: Restoring previous memory policy: 4 00:05:26.149 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.149 EAL: request: mp_malloc_sync 00:05:26.149 EAL: No shared files mode enabled, IPC is disabled 00:05:26.149 EAL: Heap on socket 0 was expanded by 514MB 00:05:26.409 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.409 EAL: request: mp_malloc_sync 00:05:26.409 EAL: No shared files mode enabled, IPC is disabled 00:05:26.409 EAL: Heap on socket 0 was shrunk by 514MB 00:05:26.409 EAL: Trying to obtain current memory policy. 00:05:26.409 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:26.669 EAL: Restoring previous memory policy: 4 00:05:26.669 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.669 EAL: request: mp_malloc_sync 00:05:26.669 EAL: No shared files mode enabled, IPC is disabled 00:05:26.669 EAL: Heap on socket 0 was expanded by 1026MB 00:05:26.928 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.928 passed 00:05:26.928 00:05:26.928 Run Summary: Type Total Ran Passed Failed Inactive 00:05:26.928 suites 1 1 n/a 0 0 00:05:26.928 tests 2 2 2 0 0 00:05:26.928 asserts 5610 5610 5610 0 n/a 00:05:26.928 00:05:26.928 Elapsed time = 1.336 seconds 00:05:26.928 EAL: request: mp_malloc_sync 00:05:26.928 EAL: No shared files mode enabled, IPC is disabled 00:05:26.928 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:26.928 EAL: Calling mem event callback 'spdk:(nil)' 00:05:26.928 EAL: request: mp_malloc_sync 00:05:26.928 EAL: No shared files mode enabled, IPC is disabled 00:05:26.928 EAL: Heap on socket 0 was shrunk by 2MB 00:05:26.928 EAL: No shared files mode enabled, IPC is disabled 00:05:26.928 EAL: No shared files mode enabled, IPC is disabled 00:05:26.928 EAL: No shared files mode enabled, IPC is disabled 00:05:26.928 00:05:26.928 real 0m1.614s 00:05:26.928 user 0m0.779s 00:05:26.928 sys 0m0.690s 00:05:26.928 14:04:31 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:26.928 14:04:31 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:26.928 ************************************ 00:05:26.928 END TEST env_vtophys 00:05:26.928 ************************************ 00:05:27.188 14:04:31 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:27.188 14:04:31 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:27.188 14:04:31 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.188 14:04:31 env -- common/autotest_common.sh@10 -- # set +x 00:05:27.188 ************************************ 00:05:27.188 START TEST env_pci 00:05:27.188 ************************************ 00:05:27.188 14:04:31 env.env_pci -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:27.188 00:05:27.188 00:05:27.188 CUnit - A unit testing framework for C - Version 2.1-3 00:05:27.188 http://cunit.sourceforge.net/ 00:05:27.188 00:05:27.188 00:05:27.188 Suite: pci 00:05:27.189 Test: pci_hook ...[2024-09-30 14:04:31.632363] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 70048 has claimed it 00:05:27.189 passed 00:05:27.189 00:05:27.189 Run Summary: Type Total Ran Passed Failed Inactive 00:05:27.189 suites 1 1 n/a 0 0 00:05:27.189 tests 1 1 1 0 0 00:05:27.189 asserts 25 25 25 0 n/a 00:05:27.189 00:05:27.189 Elapsed time = 0.007 seconds 00:05:27.189 EAL: Cannot find device (10000:00:01.0) 00:05:27.189 EAL: Failed to attach device on primary process 00:05:27.189 00:05:27.189 real 0m0.093s 00:05:27.189 user 0m0.047s 00:05:27.189 sys 0m0.044s 00:05:27.189 14:04:31 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:27.189 ************************************ 00:05:27.189 END TEST env_pci 00:05:27.189 ************************************ 00:05:27.189 14:04:31 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:27.189 14:04:31 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:27.189 14:04:31 env -- env/env.sh@15 -- # uname 00:05:27.189 14:04:31 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:27.189 14:04:31 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:27.189 14:04:31 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:27.189 14:04:31 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:05:27.189 14:04:31 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.189 14:04:31 env -- common/autotest_common.sh@10 -- # set +x 00:05:27.189 ************************************ 00:05:27.189 START TEST env_dpdk_post_init 00:05:27.189 ************************************ 00:05:27.189 14:04:31 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:27.189 EAL: Detected CPU lcores: 10 00:05:27.189 EAL: Detected NUMA nodes: 1 00:05:27.189 EAL: Detected shared linkage of DPDK 00:05:27.189 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:27.189 EAL: Selected IOVA mode 'PA' 00:05:27.449 Starting DPDK initialization... 00:05:27.449 Starting SPDK post initialization... 00:05:27.449 SPDK NVMe probe 00:05:27.449 Attaching to 0000:00:10.0 00:05:27.449 Attaching to 0000:00:11.0 00:05:27.449 Attached to 0000:00:10.0 00:05:27.449 Attached to 0000:00:11.0 00:05:27.449 Cleaning up... 00:05:27.449 ************************************ 00:05:27.449 END TEST env_dpdk_post_init 00:05:27.449 ************************************ 00:05:27.449 00:05:27.449 real 0m0.250s 00:05:27.449 user 0m0.068s 00:05:27.449 sys 0m0.083s 00:05:27.449 14:04:32 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:27.449 14:04:32 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:27.449 14:04:32 env -- env/env.sh@26 -- # uname 00:05:27.449 14:04:32 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:27.449 14:04:32 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:27.449 14:04:32 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:27.449 14:04:32 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.449 14:04:32 env -- common/autotest_common.sh@10 -- # set +x 00:05:27.449 ************************************ 00:05:27.449 START TEST env_mem_callbacks 00:05:27.449 ************************************ 00:05:27.449 14:04:32 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:27.710 EAL: Detected CPU lcores: 10 00:05:27.710 EAL: Detected NUMA nodes: 1 00:05:27.710 EAL: Detected shared linkage of DPDK 00:05:27.710 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:27.710 EAL: Selected IOVA mode 'PA' 00:05:27.710 00:05:27.710 00:05:27.710 CUnit - A unit testing framework for C - Version 2.1-3 00:05:27.710 http://cunit.sourceforge.net/ 00:05:27.710 00:05:27.710 00:05:27.710 Suite: memory 00:05:27.710 Test: test ... 00:05:27.710 register 0x200000200000 2097152 00:05:27.710 malloc 3145728 00:05:27.710 register 0x200000400000 4194304 00:05:27.710 buf 0x200000500000 len 3145728 PASSED 00:05:27.710 malloc 64 00:05:27.710 buf 0x2000004fff40 len 64 PASSED 00:05:27.710 malloc 4194304 00:05:27.710 register 0x200000800000 6291456 00:05:27.710 buf 0x200000a00000 len 4194304 PASSED 00:05:27.710 free 0x200000500000 3145728 00:05:27.710 free 0x2000004fff40 64 00:05:27.710 unregister 0x200000400000 4194304 PASSED 00:05:27.710 free 0x200000a00000 4194304 00:05:27.710 unregister 0x200000800000 6291456 PASSED 00:05:27.710 malloc 8388608 00:05:27.710 register 0x200000400000 10485760 00:05:27.710 buf 0x200000600000 len 8388608 PASSED 00:05:27.710 free 0x200000600000 8388608 00:05:27.710 unregister 0x200000400000 10485760 PASSED 00:05:27.710 passed 00:05:27.710 00:05:27.710 Run Summary: Type Total Ran Passed Failed Inactive 00:05:27.710 suites 1 1 n/a 0 0 00:05:27.710 tests 1 1 1 0 0 00:05:27.710 asserts 15 15 15 0 n/a 00:05:27.710 00:05:27.710 Elapsed time = 0.012 seconds 00:05:27.710 00:05:27.710 real 0m0.222s 00:05:27.710 user 0m0.040s 00:05:27.710 sys 0m0.079s 00:05:27.710 14:04:32 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:27.710 14:04:32 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:27.710 ************************************ 00:05:27.710 END TEST env_mem_callbacks 00:05:27.710 ************************************ 00:05:27.968 00:05:27.968 real 0m3.026s 00:05:27.968 user 0m1.400s 00:05:27.968 sys 0m1.274s 00:05:27.969 14:04:32 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:27.969 14:04:32 env -- common/autotest_common.sh@10 -- # set +x 00:05:27.969 ************************************ 00:05:27.969 END TEST env 00:05:27.969 ************************************ 00:05:27.969 14:04:32 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:27.969 14:04:32 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:27.969 14:04:32 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.969 14:04:32 -- common/autotest_common.sh@10 -- # set +x 00:05:27.969 ************************************ 00:05:27.969 START TEST rpc 00:05:27.969 ************************************ 00:05:27.969 14:04:32 rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:27.969 * Looking for test storage... 00:05:27.969 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:27.969 14:04:32 rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:27.969 14:04:32 rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:27.969 14:04:32 rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:28.229 14:04:32 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:28.229 14:04:32 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:28.229 14:04:32 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:28.229 14:04:32 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:28.229 14:04:32 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:28.229 14:04:32 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:28.229 14:04:32 rpc -- scripts/common.sh@345 -- # : 1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:28.229 14:04:32 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:28.229 14:04:32 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@353 -- # local d=1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:28.229 14:04:32 rpc -- scripts/common.sh@355 -- # echo 1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:28.229 14:04:32 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@353 -- # local d=2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:28.229 14:04:32 rpc -- scripts/common.sh@355 -- # echo 2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:28.229 14:04:32 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:28.229 14:04:32 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:28.229 14:04:32 rpc -- scripts/common.sh@368 -- # return 0 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:28.229 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:28.229 --rc genhtml_branch_coverage=1 00:05:28.229 --rc genhtml_function_coverage=1 00:05:28.229 --rc genhtml_legend=1 00:05:28.229 --rc geninfo_all_blocks=1 00:05:28.229 --rc geninfo_unexecuted_blocks=1 00:05:28.229 00:05:28.229 ' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:28.229 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:28.229 --rc genhtml_branch_coverage=1 00:05:28.229 --rc genhtml_function_coverage=1 00:05:28.229 --rc genhtml_legend=1 00:05:28.229 --rc geninfo_all_blocks=1 00:05:28.229 --rc geninfo_unexecuted_blocks=1 00:05:28.229 00:05:28.229 ' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:28.229 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:28.229 --rc genhtml_branch_coverage=1 00:05:28.229 --rc genhtml_function_coverage=1 00:05:28.229 --rc genhtml_legend=1 00:05:28.229 --rc geninfo_all_blocks=1 00:05:28.229 --rc geninfo_unexecuted_blocks=1 00:05:28.229 00:05:28.229 ' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:28.229 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:28.229 --rc genhtml_branch_coverage=1 00:05:28.229 --rc genhtml_function_coverage=1 00:05:28.229 --rc genhtml_legend=1 00:05:28.229 --rc geninfo_all_blocks=1 00:05:28.229 --rc geninfo_unexecuted_blocks=1 00:05:28.229 00:05:28.229 ' 00:05:28.229 14:04:32 rpc -- rpc/rpc.sh@65 -- # spdk_pid=70175 00:05:28.229 14:04:32 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:28.229 14:04:32 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:28.229 14:04:32 rpc -- rpc/rpc.sh@67 -- # waitforlisten 70175 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@831 -- # '[' -z 70175 ']' 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:28.229 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:28.229 14:04:32 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:28.229 [2024-09-30 14:04:32.783648] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:28.229 [2024-09-30 14:04:32.783798] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70175 ] 00:05:28.489 [2024-09-30 14:04:32.922515] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:28.489 [2024-09-30 14:04:32.953071] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:28.489 [2024-09-30 14:04:33.009644] app.c: 610:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:28.489 [2024-09-30 14:04:33.009723] app.c: 611:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 70175' to capture a snapshot of events at runtime. 00:05:28.489 [2024-09-30 14:04:33.009734] app.c: 616:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:28.489 [2024-09-30 14:04:33.009759] app.c: 617:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:28.489 [2024-09-30 14:04:33.009767] app.c: 618:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid70175 for offline analysis/debug. 00:05:28.489 [2024-09-30 14:04:33.009863] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:29.058 14:04:33 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:29.058 14:04:33 rpc -- common/autotest_common.sh@864 -- # return 0 00:05:29.058 14:04:33 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:29.058 14:04:33 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:29.058 14:04:33 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:29.058 14:04:33 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:29.058 14:04:33 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:29.058 14:04:33 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:29.058 14:04:33 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:29.058 ************************************ 00:05:29.058 START TEST rpc_integrity 00:05:29.058 ************************************ 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.058 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:29.058 { 00:05:29.058 "name": "Malloc0", 00:05:29.058 "aliases": [ 00:05:29.058 "9a5d2d58-b11b-4c92-9e2b-b1812373becb" 00:05:29.058 ], 00:05:29.058 "product_name": "Malloc disk", 00:05:29.058 "block_size": 512, 00:05:29.058 "num_blocks": 16384, 00:05:29.058 "uuid": "9a5d2d58-b11b-4c92-9e2b-b1812373becb", 00:05:29.058 "assigned_rate_limits": { 00:05:29.058 "rw_ios_per_sec": 0, 00:05:29.058 "rw_mbytes_per_sec": 0, 00:05:29.058 "r_mbytes_per_sec": 0, 00:05:29.058 "w_mbytes_per_sec": 0 00:05:29.058 }, 00:05:29.058 "claimed": false, 00:05:29.058 "zoned": false, 00:05:29.058 "supported_io_types": { 00:05:29.058 "read": true, 00:05:29.058 "write": true, 00:05:29.058 "unmap": true, 00:05:29.058 "flush": true, 00:05:29.058 "reset": true, 00:05:29.058 "nvme_admin": false, 00:05:29.058 "nvme_io": false, 00:05:29.058 "nvme_io_md": false, 00:05:29.058 "write_zeroes": true, 00:05:29.058 "zcopy": true, 00:05:29.058 "get_zone_info": false, 00:05:29.058 "zone_management": false, 00:05:29.058 "zone_append": false, 00:05:29.058 "compare": false, 00:05:29.058 "compare_and_write": false, 00:05:29.058 "abort": true, 00:05:29.058 "seek_hole": false, 00:05:29.058 "seek_data": false, 00:05:29.058 "copy": true, 00:05:29.058 "nvme_iov_md": false 00:05:29.058 }, 00:05:29.058 "memory_domains": [ 00:05:29.058 { 00:05:29.058 "dma_device_id": "system", 00:05:29.058 "dma_device_type": 1 00:05:29.058 }, 00:05:29.058 { 00:05:29.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:29.058 "dma_device_type": 2 00:05:29.058 } 00:05:29.058 ], 00:05:29.058 "driver_specific": {} 00:05:29.058 } 00:05:29.058 ]' 00:05:29.058 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 [2024-09-30 14:04:33.747226] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:29.318 [2024-09-30 14:04:33.747285] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:29.318 [2024-09-30 14:04:33.747329] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:05:29.318 [2024-09-30 14:04:33.747348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:29.318 [2024-09-30 14:04:33.749626] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:29.318 [2024-09-30 14:04:33.749662] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:29.318 Passthru0 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:29.318 { 00:05:29.318 "name": "Malloc0", 00:05:29.318 "aliases": [ 00:05:29.318 "9a5d2d58-b11b-4c92-9e2b-b1812373becb" 00:05:29.318 ], 00:05:29.318 "product_name": "Malloc disk", 00:05:29.318 "block_size": 512, 00:05:29.318 "num_blocks": 16384, 00:05:29.318 "uuid": "9a5d2d58-b11b-4c92-9e2b-b1812373becb", 00:05:29.318 "assigned_rate_limits": { 00:05:29.318 "rw_ios_per_sec": 0, 00:05:29.318 "rw_mbytes_per_sec": 0, 00:05:29.318 "r_mbytes_per_sec": 0, 00:05:29.318 "w_mbytes_per_sec": 0 00:05:29.318 }, 00:05:29.318 "claimed": true, 00:05:29.318 "claim_type": "exclusive_write", 00:05:29.318 "zoned": false, 00:05:29.318 "supported_io_types": { 00:05:29.318 "read": true, 00:05:29.318 "write": true, 00:05:29.318 "unmap": true, 00:05:29.318 "flush": true, 00:05:29.318 "reset": true, 00:05:29.318 "nvme_admin": false, 00:05:29.318 "nvme_io": false, 00:05:29.318 "nvme_io_md": false, 00:05:29.318 "write_zeroes": true, 00:05:29.318 "zcopy": true, 00:05:29.318 "get_zone_info": false, 00:05:29.318 "zone_management": false, 00:05:29.318 "zone_append": false, 00:05:29.318 "compare": false, 00:05:29.318 "compare_and_write": false, 00:05:29.318 "abort": true, 00:05:29.318 "seek_hole": false, 00:05:29.318 "seek_data": false, 00:05:29.318 "copy": true, 00:05:29.318 "nvme_iov_md": false 00:05:29.318 }, 00:05:29.318 "memory_domains": [ 00:05:29.318 { 00:05:29.318 "dma_device_id": "system", 00:05:29.318 "dma_device_type": 1 00:05:29.318 }, 00:05:29.318 { 00:05:29.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:29.318 "dma_device_type": 2 00:05:29.318 } 00:05:29.318 ], 00:05:29.318 "driver_specific": {} 00:05:29.318 }, 00:05:29.318 { 00:05:29.318 "name": "Passthru0", 00:05:29.318 "aliases": [ 00:05:29.318 "aad13d71-f6d1-501e-8b5a-a778fd0cd467" 00:05:29.318 ], 00:05:29.318 "product_name": "passthru", 00:05:29.318 "block_size": 512, 00:05:29.318 "num_blocks": 16384, 00:05:29.318 "uuid": "aad13d71-f6d1-501e-8b5a-a778fd0cd467", 00:05:29.318 "assigned_rate_limits": { 00:05:29.318 "rw_ios_per_sec": 0, 00:05:29.318 "rw_mbytes_per_sec": 0, 00:05:29.318 "r_mbytes_per_sec": 0, 00:05:29.318 "w_mbytes_per_sec": 0 00:05:29.318 }, 00:05:29.318 "claimed": false, 00:05:29.318 "zoned": false, 00:05:29.318 "supported_io_types": { 00:05:29.318 "read": true, 00:05:29.318 "write": true, 00:05:29.318 "unmap": true, 00:05:29.318 "flush": true, 00:05:29.318 "reset": true, 00:05:29.318 "nvme_admin": false, 00:05:29.318 "nvme_io": false, 00:05:29.318 "nvme_io_md": false, 00:05:29.318 "write_zeroes": true, 00:05:29.318 "zcopy": true, 00:05:29.318 "get_zone_info": false, 00:05:29.318 "zone_management": false, 00:05:29.318 "zone_append": false, 00:05:29.318 "compare": false, 00:05:29.318 "compare_and_write": false, 00:05:29.318 "abort": true, 00:05:29.318 "seek_hole": false, 00:05:29.318 "seek_data": false, 00:05:29.318 "copy": true, 00:05:29.318 "nvme_iov_md": false 00:05:29.318 }, 00:05:29.318 "memory_domains": [ 00:05:29.318 { 00:05:29.318 "dma_device_id": "system", 00:05:29.318 "dma_device_type": 1 00:05:29.318 }, 00:05:29.318 { 00:05:29.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:29.318 "dma_device_type": 2 00:05:29.318 } 00:05:29.318 ], 00:05:29.318 "driver_specific": { 00:05:29.318 "passthru": { 00:05:29.318 "name": "Passthru0", 00:05:29.318 "base_bdev_name": "Malloc0" 00:05:29.318 } 00:05:29.318 } 00:05:29.318 } 00:05:29.318 ]' 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:29.318 ************************************ 00:05:29.318 END TEST rpc_integrity 00:05:29.318 ************************************ 00:05:29.318 14:04:33 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:29.318 00:05:29.318 real 0m0.309s 00:05:29.318 user 0m0.177s 00:05:29.318 sys 0m0.055s 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:29.318 14:04:33 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:29.318 14:04:33 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:29.318 14:04:33 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 ************************************ 00:05:29.318 START TEST rpc_plugins 00:05:29.318 ************************************ 00:05:29.318 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:05:29.318 14:04:33 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:29.318 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.318 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:29.318 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.577 14:04:33 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:29.577 14:04:33 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:29.577 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.577 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:29.577 14:04:33 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.577 14:04:33 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:29.577 { 00:05:29.577 "name": "Malloc1", 00:05:29.577 "aliases": [ 00:05:29.577 "b07f09ad-1e65-43a5-ace9-b7725280c3c1" 00:05:29.577 ], 00:05:29.577 "product_name": "Malloc disk", 00:05:29.577 "block_size": 4096, 00:05:29.577 "num_blocks": 256, 00:05:29.577 "uuid": "b07f09ad-1e65-43a5-ace9-b7725280c3c1", 00:05:29.577 "assigned_rate_limits": { 00:05:29.577 "rw_ios_per_sec": 0, 00:05:29.577 "rw_mbytes_per_sec": 0, 00:05:29.577 "r_mbytes_per_sec": 0, 00:05:29.577 "w_mbytes_per_sec": 0 00:05:29.577 }, 00:05:29.577 "claimed": false, 00:05:29.577 "zoned": false, 00:05:29.577 "supported_io_types": { 00:05:29.577 "read": true, 00:05:29.577 "write": true, 00:05:29.577 "unmap": true, 00:05:29.578 "flush": true, 00:05:29.578 "reset": true, 00:05:29.578 "nvme_admin": false, 00:05:29.578 "nvme_io": false, 00:05:29.578 "nvme_io_md": false, 00:05:29.578 "write_zeroes": true, 00:05:29.578 "zcopy": true, 00:05:29.578 "get_zone_info": false, 00:05:29.578 "zone_management": false, 00:05:29.578 "zone_append": false, 00:05:29.578 "compare": false, 00:05:29.578 "compare_and_write": false, 00:05:29.578 "abort": true, 00:05:29.578 "seek_hole": false, 00:05:29.578 "seek_data": false, 00:05:29.578 "copy": true, 00:05:29.578 "nvme_iov_md": false 00:05:29.578 }, 00:05:29.578 "memory_domains": [ 00:05:29.578 { 00:05:29.578 "dma_device_id": "system", 00:05:29.578 "dma_device_type": 1 00:05:29.578 }, 00:05:29.578 { 00:05:29.578 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:29.578 "dma_device_type": 2 00:05:29.578 } 00:05:29.578 ], 00:05:29.578 "driver_specific": {} 00:05:29.578 } 00:05:29.578 ]' 00:05:29.578 14:04:33 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:29.578 ************************************ 00:05:29.578 END TEST rpc_plugins 00:05:29.578 ************************************ 00:05:29.578 14:04:34 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:29.578 00:05:29.578 real 0m0.145s 00:05:29.578 user 0m0.085s 00:05:29.578 sys 0m0.021s 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:29.578 14:04:34 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:29.578 14:04:34 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:29.578 14:04:34 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:29.578 14:04:34 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:29.578 14:04:34 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:29.578 ************************************ 00:05:29.578 START TEST rpc_trace_cmd_test 00:05:29.578 ************************************ 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:29.578 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid70175", 00:05:29.578 "tpoint_group_mask": "0x8", 00:05:29.578 "iscsi_conn": { 00:05:29.578 "mask": "0x2", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "scsi": { 00:05:29.578 "mask": "0x4", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "bdev": { 00:05:29.578 "mask": "0x8", 00:05:29.578 "tpoint_mask": "0xffffffffffffffff" 00:05:29.578 }, 00:05:29.578 "nvmf_rdma": { 00:05:29.578 "mask": "0x10", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "nvmf_tcp": { 00:05:29.578 "mask": "0x20", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "ftl": { 00:05:29.578 "mask": "0x40", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "blobfs": { 00:05:29.578 "mask": "0x80", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "dsa": { 00:05:29.578 "mask": "0x200", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "thread": { 00:05:29.578 "mask": "0x400", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "nvme_pcie": { 00:05:29.578 "mask": "0x800", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "iaa": { 00:05:29.578 "mask": "0x1000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "nvme_tcp": { 00:05:29.578 "mask": "0x2000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "bdev_nvme": { 00:05:29.578 "mask": "0x4000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "sock": { 00:05:29.578 "mask": "0x8000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "blob": { 00:05:29.578 "mask": "0x10000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 }, 00:05:29.578 "bdev_raid": { 00:05:29.578 "mask": "0x20000", 00:05:29.578 "tpoint_mask": "0x0" 00:05:29.578 } 00:05:29.578 }' 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 18 -gt 2 ']' 00:05:29.578 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:29.837 ************************************ 00:05:29.837 END TEST rpc_trace_cmd_test 00:05:29.837 ************************************ 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:29.837 00:05:29.837 real 0m0.218s 00:05:29.837 user 0m0.179s 00:05:29.837 sys 0m0.027s 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:29.837 14:04:34 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:29.837 14:04:34 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:29.837 14:04:34 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:29.837 14:04:34 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:29.837 14:04:34 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:29.837 14:04:34 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:29.837 14:04:34 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:29.837 ************************************ 00:05:29.837 START TEST rpc_daemon_integrity 00:05:29.837 ************************************ 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:29.837 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:30.096 { 00:05:30.096 "name": "Malloc2", 00:05:30.096 "aliases": [ 00:05:30.096 "29ed824f-c092-4ee3-b654-a05cd2eb5363" 00:05:30.096 ], 00:05:30.096 "product_name": "Malloc disk", 00:05:30.096 "block_size": 512, 00:05:30.096 "num_blocks": 16384, 00:05:30.096 "uuid": "29ed824f-c092-4ee3-b654-a05cd2eb5363", 00:05:30.096 "assigned_rate_limits": { 00:05:30.096 "rw_ios_per_sec": 0, 00:05:30.096 "rw_mbytes_per_sec": 0, 00:05:30.096 "r_mbytes_per_sec": 0, 00:05:30.096 "w_mbytes_per_sec": 0 00:05:30.096 }, 00:05:30.096 "claimed": false, 00:05:30.096 "zoned": false, 00:05:30.096 "supported_io_types": { 00:05:30.096 "read": true, 00:05:30.096 "write": true, 00:05:30.096 "unmap": true, 00:05:30.096 "flush": true, 00:05:30.096 "reset": true, 00:05:30.096 "nvme_admin": false, 00:05:30.096 "nvme_io": false, 00:05:30.096 "nvme_io_md": false, 00:05:30.096 "write_zeroes": true, 00:05:30.096 "zcopy": true, 00:05:30.096 "get_zone_info": false, 00:05:30.096 "zone_management": false, 00:05:30.096 "zone_append": false, 00:05:30.096 "compare": false, 00:05:30.096 "compare_and_write": false, 00:05:30.096 "abort": true, 00:05:30.096 "seek_hole": false, 00:05:30.096 "seek_data": false, 00:05:30.096 "copy": true, 00:05:30.096 "nvme_iov_md": false 00:05:30.096 }, 00:05:30.096 "memory_domains": [ 00:05:30.096 { 00:05:30.096 "dma_device_id": "system", 00:05:30.096 "dma_device_type": 1 00:05:30.096 }, 00:05:30.096 { 00:05:30.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:30.096 "dma_device_type": 2 00:05:30.096 } 00:05:30.096 ], 00:05:30.096 "driver_specific": {} 00:05:30.096 } 00:05:30.096 ]' 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 [2024-09-30 14:04:34.578646] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:30.096 [2024-09-30 14:04:34.578706] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:30.096 [2024-09-30 14:04:34.578726] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:05:30.096 [2024-09-30 14:04:34.578738] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:30.096 [2024-09-30 14:04:34.580922] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:30.096 [2024-09-30 14:04:34.580961] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:30.096 Passthru0 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:30.096 { 00:05:30.096 "name": "Malloc2", 00:05:30.096 "aliases": [ 00:05:30.096 "29ed824f-c092-4ee3-b654-a05cd2eb5363" 00:05:30.096 ], 00:05:30.096 "product_name": "Malloc disk", 00:05:30.096 "block_size": 512, 00:05:30.096 "num_blocks": 16384, 00:05:30.096 "uuid": "29ed824f-c092-4ee3-b654-a05cd2eb5363", 00:05:30.096 "assigned_rate_limits": { 00:05:30.096 "rw_ios_per_sec": 0, 00:05:30.096 "rw_mbytes_per_sec": 0, 00:05:30.096 "r_mbytes_per_sec": 0, 00:05:30.096 "w_mbytes_per_sec": 0 00:05:30.096 }, 00:05:30.096 "claimed": true, 00:05:30.096 "claim_type": "exclusive_write", 00:05:30.096 "zoned": false, 00:05:30.096 "supported_io_types": { 00:05:30.096 "read": true, 00:05:30.096 "write": true, 00:05:30.096 "unmap": true, 00:05:30.096 "flush": true, 00:05:30.096 "reset": true, 00:05:30.096 "nvme_admin": false, 00:05:30.096 "nvme_io": false, 00:05:30.096 "nvme_io_md": false, 00:05:30.096 "write_zeroes": true, 00:05:30.096 "zcopy": true, 00:05:30.096 "get_zone_info": false, 00:05:30.096 "zone_management": false, 00:05:30.096 "zone_append": false, 00:05:30.096 "compare": false, 00:05:30.096 "compare_and_write": false, 00:05:30.096 "abort": true, 00:05:30.096 "seek_hole": false, 00:05:30.096 "seek_data": false, 00:05:30.096 "copy": true, 00:05:30.096 "nvme_iov_md": false 00:05:30.096 }, 00:05:30.096 "memory_domains": [ 00:05:30.096 { 00:05:30.096 "dma_device_id": "system", 00:05:30.096 "dma_device_type": 1 00:05:30.096 }, 00:05:30.096 { 00:05:30.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:30.096 "dma_device_type": 2 00:05:30.096 } 00:05:30.096 ], 00:05:30.096 "driver_specific": {} 00:05:30.096 }, 00:05:30.096 { 00:05:30.096 "name": "Passthru0", 00:05:30.096 "aliases": [ 00:05:30.096 "47ad37d8-d3fa-5d5a-b96c-8a362ef39e8a" 00:05:30.096 ], 00:05:30.096 "product_name": "passthru", 00:05:30.096 "block_size": 512, 00:05:30.096 "num_blocks": 16384, 00:05:30.096 "uuid": "47ad37d8-d3fa-5d5a-b96c-8a362ef39e8a", 00:05:30.096 "assigned_rate_limits": { 00:05:30.096 "rw_ios_per_sec": 0, 00:05:30.096 "rw_mbytes_per_sec": 0, 00:05:30.096 "r_mbytes_per_sec": 0, 00:05:30.096 "w_mbytes_per_sec": 0 00:05:30.096 }, 00:05:30.096 "claimed": false, 00:05:30.096 "zoned": false, 00:05:30.096 "supported_io_types": { 00:05:30.096 "read": true, 00:05:30.096 "write": true, 00:05:30.096 "unmap": true, 00:05:30.096 "flush": true, 00:05:30.096 "reset": true, 00:05:30.096 "nvme_admin": false, 00:05:30.096 "nvme_io": false, 00:05:30.096 "nvme_io_md": false, 00:05:30.096 "write_zeroes": true, 00:05:30.096 "zcopy": true, 00:05:30.096 "get_zone_info": false, 00:05:30.096 "zone_management": false, 00:05:30.096 "zone_append": false, 00:05:30.096 "compare": false, 00:05:30.096 "compare_and_write": false, 00:05:30.096 "abort": true, 00:05:30.096 "seek_hole": false, 00:05:30.096 "seek_data": false, 00:05:30.096 "copy": true, 00:05:30.096 "nvme_iov_md": false 00:05:30.096 }, 00:05:30.096 "memory_domains": [ 00:05:30.096 { 00:05:30.096 "dma_device_id": "system", 00:05:30.096 "dma_device_type": 1 00:05:30.096 }, 00:05:30.096 { 00:05:30.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:30.096 "dma_device_type": 2 00:05:30.096 } 00:05:30.096 ], 00:05:30.096 "driver_specific": { 00:05:30.096 "passthru": { 00:05:30.096 "name": "Passthru0", 00:05:30.096 "base_bdev_name": "Malloc2" 00:05:30.096 } 00:05:30.096 } 00:05:30.096 } 00:05:30.096 ]' 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.096 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:30.097 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:30.097 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:30.356 ************************************ 00:05:30.356 END TEST rpc_daemon_integrity 00:05:30.356 ************************************ 00:05:30.356 14:04:34 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:30.356 00:05:30.356 real 0m0.316s 00:05:30.356 user 0m0.186s 00:05:30.356 sys 0m0.057s 00:05:30.356 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:30.356 14:04:34 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:30.356 14:04:34 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:30.356 14:04:34 rpc -- rpc/rpc.sh@84 -- # killprocess 70175 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@950 -- # '[' -z 70175 ']' 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@954 -- # kill -0 70175 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@955 -- # uname 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70175 00:05:30.356 killing process with pid 70175 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70175' 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@969 -- # kill 70175 00:05:30.356 14:04:34 rpc -- common/autotest_common.sh@974 -- # wait 70175 00:05:30.616 00:05:30.616 real 0m2.803s 00:05:30.616 user 0m3.263s 00:05:30.616 sys 0m0.876s 00:05:30.616 14:04:35 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:30.616 14:04:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:30.616 ************************************ 00:05:30.616 END TEST rpc 00:05:30.616 ************************************ 00:05:30.875 14:04:35 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:30.875 14:04:35 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:30.875 14:04:35 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:30.875 14:04:35 -- common/autotest_common.sh@10 -- # set +x 00:05:30.875 ************************************ 00:05:30.875 START TEST skip_rpc 00:05:30.875 ************************************ 00:05:30.875 14:04:35 skip_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:30.875 * Looking for test storage... 00:05:30.875 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:30.875 14:04:35 skip_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:30.875 14:04:35 skip_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:30.875 14:04:35 skip_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:30.875 14:04:35 skip_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:30.875 14:04:35 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:31.135 14:04:35 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:31.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:31.135 --rc genhtml_branch_coverage=1 00:05:31.135 --rc genhtml_function_coverage=1 00:05:31.135 --rc genhtml_legend=1 00:05:31.135 --rc geninfo_all_blocks=1 00:05:31.135 --rc geninfo_unexecuted_blocks=1 00:05:31.135 00:05:31.135 ' 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:31.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:31.135 --rc genhtml_branch_coverage=1 00:05:31.135 --rc genhtml_function_coverage=1 00:05:31.135 --rc genhtml_legend=1 00:05:31.135 --rc geninfo_all_blocks=1 00:05:31.135 --rc geninfo_unexecuted_blocks=1 00:05:31.135 00:05:31.135 ' 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:31.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:31.135 --rc genhtml_branch_coverage=1 00:05:31.135 --rc genhtml_function_coverage=1 00:05:31.135 --rc genhtml_legend=1 00:05:31.135 --rc geninfo_all_blocks=1 00:05:31.135 --rc geninfo_unexecuted_blocks=1 00:05:31.135 00:05:31.135 ' 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:31.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:31.135 --rc genhtml_branch_coverage=1 00:05:31.135 --rc genhtml_function_coverage=1 00:05:31.135 --rc genhtml_legend=1 00:05:31.135 --rc geninfo_all_blocks=1 00:05:31.135 --rc geninfo_unexecuted_blocks=1 00:05:31.135 00:05:31.135 ' 00:05:31.135 14:04:35 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:31.135 14:04:35 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:31.135 14:04:35 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:31.135 14:04:35 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:31.135 ************************************ 00:05:31.135 START TEST skip_rpc 00:05:31.135 ************************************ 00:05:31.135 14:04:35 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:05:31.135 14:04:35 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=70379 00:05:31.135 14:04:35 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:31.135 14:04:35 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:31.135 14:04:35 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:31.135 [2024-09-30 14:04:35.661799] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:31.135 [2024-09-30 14:04:35.661926] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70379 ] 00:05:31.395 [2024-09-30 14:04:35.792919] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:31.395 [2024-09-30 14:04:35.820360] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:31.395 [2024-09-30 14:04:35.878801] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 70379 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 70379 ']' 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 70379 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70379 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:36.672 killing process with pid 70379 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70379' 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 70379 00:05:36.672 14:04:40 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 70379 00:05:36.672 00:05:36.672 real 0m5.447s 00:05:36.672 user 0m5.031s 00:05:36.672 sys 0m0.334s 00:05:36.672 14:04:41 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:36.672 ************************************ 00:05:36.672 END TEST skip_rpc 00:05:36.672 ************************************ 00:05:36.672 14:04:41 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:36.672 14:04:41 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:36.672 14:04:41 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:36.672 14:04:41 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:36.672 14:04:41 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:36.672 ************************************ 00:05:36.672 START TEST skip_rpc_with_json 00:05:36.672 ************************************ 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=70464 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 70464 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 70464 ']' 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:36.672 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:36.672 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:36.672 [2024-09-30 14:04:41.181750] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:36.672 [2024-09-30 14:04:41.181884] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70464 ] 00:05:36.672 [2024-09-30 14:04:41.323948] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:36.933 [2024-09-30 14:04:41.352408] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:36.933 [2024-09-30 14:04:41.398611] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:37.501 [2024-09-30 14:04:41.986400] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:37.501 request: 00:05:37.501 { 00:05:37.501 "trtype": "tcp", 00:05:37.501 "method": "nvmf_get_transports", 00:05:37.501 "req_id": 1 00:05:37.501 } 00:05:37.501 Got JSON-RPC error response 00:05:37.501 response: 00:05:37.501 { 00:05:37.501 "code": -19, 00:05:37.501 "message": "No such device" 00:05:37.501 } 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:37.501 14:04:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:37.501 [2024-09-30 14:04:41.998533] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:37.501 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:37.501 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:37.501 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:37.502 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:37.762 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:37.762 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:37.762 { 00:05:37.762 "subsystems": [ 00:05:37.762 { 00:05:37.762 "subsystem": "fsdev", 00:05:37.762 "config": [ 00:05:37.762 { 00:05:37.762 "method": "fsdev_set_opts", 00:05:37.762 "params": { 00:05:37.762 "fsdev_io_pool_size": 65535, 00:05:37.762 "fsdev_io_cache_size": 256 00:05:37.762 } 00:05:37.762 } 00:05:37.762 ] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "keyring", 00:05:37.762 "config": [] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "iobuf", 00:05:37.762 "config": [ 00:05:37.762 { 00:05:37.762 "method": "iobuf_set_options", 00:05:37.762 "params": { 00:05:37.762 "small_pool_count": 8192, 00:05:37.762 "large_pool_count": 1024, 00:05:37.762 "small_bufsize": 8192, 00:05:37.762 "large_bufsize": 135168 00:05:37.762 } 00:05:37.762 } 00:05:37.762 ] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "sock", 00:05:37.762 "config": [ 00:05:37.762 { 00:05:37.762 "method": "sock_set_default_impl", 00:05:37.762 "params": { 00:05:37.762 "impl_name": "posix" 00:05:37.762 } 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "method": "sock_impl_set_options", 00:05:37.762 "params": { 00:05:37.762 "impl_name": "ssl", 00:05:37.762 "recv_buf_size": 4096, 00:05:37.762 "send_buf_size": 4096, 00:05:37.762 "enable_recv_pipe": true, 00:05:37.762 "enable_quickack": false, 00:05:37.762 "enable_placement_id": 0, 00:05:37.762 "enable_zerocopy_send_server": true, 00:05:37.762 "enable_zerocopy_send_client": false, 00:05:37.762 "zerocopy_threshold": 0, 00:05:37.762 "tls_version": 0, 00:05:37.762 "enable_ktls": false 00:05:37.762 } 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "method": "sock_impl_set_options", 00:05:37.762 "params": { 00:05:37.762 "impl_name": "posix", 00:05:37.762 "recv_buf_size": 2097152, 00:05:37.762 "send_buf_size": 2097152, 00:05:37.762 "enable_recv_pipe": true, 00:05:37.762 "enable_quickack": false, 00:05:37.762 "enable_placement_id": 0, 00:05:37.762 "enable_zerocopy_send_server": true, 00:05:37.762 "enable_zerocopy_send_client": false, 00:05:37.762 "zerocopy_threshold": 0, 00:05:37.762 "tls_version": 0, 00:05:37.762 "enable_ktls": false 00:05:37.762 } 00:05:37.762 } 00:05:37.762 ] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "vmd", 00:05:37.762 "config": [] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "accel", 00:05:37.762 "config": [ 00:05:37.762 { 00:05:37.762 "method": "accel_set_options", 00:05:37.762 "params": { 00:05:37.762 "small_cache_size": 128, 00:05:37.762 "large_cache_size": 16, 00:05:37.762 "task_count": 2048, 00:05:37.762 "sequence_count": 2048, 00:05:37.762 "buf_count": 2048 00:05:37.762 } 00:05:37.762 } 00:05:37.762 ] 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "subsystem": "bdev", 00:05:37.762 "config": [ 00:05:37.762 { 00:05:37.762 "method": "bdev_set_options", 00:05:37.762 "params": { 00:05:37.762 "bdev_io_pool_size": 65535, 00:05:37.762 "bdev_io_cache_size": 256, 00:05:37.762 "bdev_auto_examine": true, 00:05:37.762 "iobuf_small_cache_size": 128, 00:05:37.762 "iobuf_large_cache_size": 16 00:05:37.762 } 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "method": "bdev_raid_set_options", 00:05:37.762 "params": { 00:05:37.762 "process_window_size_kb": 1024, 00:05:37.762 "process_max_bandwidth_mb_sec": 0 00:05:37.762 } 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "method": "bdev_iscsi_set_options", 00:05:37.762 "params": { 00:05:37.762 "timeout_sec": 30 00:05:37.762 } 00:05:37.762 }, 00:05:37.762 { 00:05:37.762 "method": "bdev_nvme_set_options", 00:05:37.762 "params": { 00:05:37.762 "action_on_timeout": "none", 00:05:37.762 "timeout_us": 0, 00:05:37.762 "timeout_admin_us": 0, 00:05:37.762 "keep_alive_timeout_ms": 10000, 00:05:37.763 "arbitration_burst": 0, 00:05:37.763 "low_priority_weight": 0, 00:05:37.763 "medium_priority_weight": 0, 00:05:37.763 "high_priority_weight": 0, 00:05:37.763 "nvme_adminq_poll_period_us": 10000, 00:05:37.763 "nvme_ioq_poll_period_us": 0, 00:05:37.763 "io_queue_requests": 0, 00:05:37.763 "delay_cmd_submit": true, 00:05:37.763 "transport_retry_count": 4, 00:05:37.763 "bdev_retry_count": 3, 00:05:37.763 "transport_ack_timeout": 0, 00:05:37.763 "ctrlr_loss_timeout_sec": 0, 00:05:37.763 "reconnect_delay_sec": 0, 00:05:37.763 "fast_io_fail_timeout_sec": 0, 00:05:37.763 "disable_auto_failback": false, 00:05:37.763 "generate_uuids": false, 00:05:37.763 "transport_tos": 0, 00:05:37.763 "nvme_error_stat": false, 00:05:37.763 "rdma_srq_size": 0, 00:05:37.763 "io_path_stat": false, 00:05:37.763 "allow_accel_sequence": false, 00:05:37.763 "rdma_max_cq_size": 0, 00:05:37.763 "rdma_cm_event_timeout_ms": 0, 00:05:37.763 "dhchap_digests": [ 00:05:37.763 "sha256", 00:05:37.763 "sha384", 00:05:37.763 "sha512" 00:05:37.763 ], 00:05:37.763 "dhchap_dhgroups": [ 00:05:37.763 "null", 00:05:37.763 "ffdhe2048", 00:05:37.763 "ffdhe3072", 00:05:37.763 "ffdhe4096", 00:05:37.763 "ffdhe6144", 00:05:37.763 "ffdhe8192" 00:05:37.763 ] 00:05:37.763 } 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "method": "bdev_nvme_set_hotplug", 00:05:37.763 "params": { 00:05:37.763 "period_us": 100000, 00:05:37.763 "enable": false 00:05:37.763 } 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "method": "bdev_wait_for_examine" 00:05:37.763 } 00:05:37.763 ] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "scsi", 00:05:37.763 "config": null 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "scheduler", 00:05:37.763 "config": [ 00:05:37.763 { 00:05:37.763 "method": "framework_set_scheduler", 00:05:37.763 "params": { 00:05:37.763 "name": "static" 00:05:37.763 } 00:05:37.763 } 00:05:37.763 ] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "vhost_scsi", 00:05:37.763 "config": [] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "vhost_blk", 00:05:37.763 "config": [] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "ublk", 00:05:37.763 "config": [] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "nbd", 00:05:37.763 "config": [] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "nvmf", 00:05:37.763 "config": [ 00:05:37.763 { 00:05:37.763 "method": "nvmf_set_config", 00:05:37.763 "params": { 00:05:37.763 "discovery_filter": "match_any", 00:05:37.763 "admin_cmd_passthru": { 00:05:37.763 "identify_ctrlr": false 00:05:37.763 }, 00:05:37.763 "dhchap_digests": [ 00:05:37.763 "sha256", 00:05:37.763 "sha384", 00:05:37.763 "sha512" 00:05:37.763 ], 00:05:37.763 "dhchap_dhgroups": [ 00:05:37.763 "null", 00:05:37.763 "ffdhe2048", 00:05:37.763 "ffdhe3072", 00:05:37.763 "ffdhe4096", 00:05:37.763 "ffdhe6144", 00:05:37.763 "ffdhe8192" 00:05:37.763 ] 00:05:37.763 } 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "method": "nvmf_set_max_subsystems", 00:05:37.763 "params": { 00:05:37.763 "max_subsystems": 1024 00:05:37.763 } 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "method": "nvmf_set_crdt", 00:05:37.763 "params": { 00:05:37.763 "crdt1": 0, 00:05:37.763 "crdt2": 0, 00:05:37.763 "crdt3": 0 00:05:37.763 } 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "method": "nvmf_create_transport", 00:05:37.763 "params": { 00:05:37.763 "trtype": "TCP", 00:05:37.763 "max_queue_depth": 128, 00:05:37.763 "max_io_qpairs_per_ctrlr": 127, 00:05:37.763 "in_capsule_data_size": 4096, 00:05:37.763 "max_io_size": 131072, 00:05:37.763 "io_unit_size": 131072, 00:05:37.763 "max_aq_depth": 128, 00:05:37.763 "num_shared_buffers": 511, 00:05:37.763 "buf_cache_size": 4294967295, 00:05:37.763 "dif_insert_or_strip": false, 00:05:37.763 "zcopy": false, 00:05:37.763 "c2h_success": true, 00:05:37.763 "sock_priority": 0, 00:05:37.763 "abort_timeout_sec": 1, 00:05:37.763 "ack_timeout": 0, 00:05:37.763 "data_wr_pool_size": 0 00:05:37.763 } 00:05:37.763 } 00:05:37.763 ] 00:05:37.763 }, 00:05:37.763 { 00:05:37.763 "subsystem": "iscsi", 00:05:37.763 "config": [ 00:05:37.763 { 00:05:37.763 "method": "iscsi_set_options", 00:05:37.763 "params": { 00:05:37.763 "node_base": "iqn.2016-06.io.spdk", 00:05:37.763 "max_sessions": 128, 00:05:37.763 "max_connections_per_session": 2, 00:05:37.763 "max_queue_depth": 64, 00:05:37.763 "default_time2wait": 2, 00:05:37.763 "default_time2retain": 20, 00:05:37.763 "first_burst_length": 8192, 00:05:37.763 "immediate_data": true, 00:05:37.763 "allow_duplicated_isid": false, 00:05:37.763 "error_recovery_level": 0, 00:05:37.763 "nop_timeout": 60, 00:05:37.763 "nop_in_interval": 30, 00:05:37.763 "disable_chap": false, 00:05:37.763 "require_chap": false, 00:05:37.763 "mutual_chap": false, 00:05:37.763 "chap_group": 0, 00:05:37.763 "max_large_datain_per_connection": 64, 00:05:37.763 "max_r2t_per_connection": 4, 00:05:37.763 "pdu_pool_size": 36864, 00:05:37.763 "immediate_data_pool_size": 16384, 00:05:37.763 "data_out_pool_size": 2048 00:05:37.763 } 00:05:37.763 } 00:05:37.763 ] 00:05:37.763 } 00:05:37.763 ] 00:05:37.763 } 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 70464 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 70464 ']' 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 70464 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70464 00:05:37.763 killing process with pid 70464 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70464' 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 70464 00:05:37.763 14:04:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 70464 00:05:38.024 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=70493 00:05:38.024 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:38.024 14:04:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 70493 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 70493 ']' 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 70493 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70493 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:43.349 killing process with pid 70493 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70493' 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 70493 00:05:43.349 14:04:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 70493 00:05:43.608 14:04:48 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:43.608 14:04:48 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:43.608 00:05:43.608 real 0m6.953s 00:05:43.608 user 0m6.436s 00:05:43.608 sys 0m0.768s 00:05:43.608 14:04:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:43.608 14:04:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:43.608 ************************************ 00:05:43.608 END TEST skip_rpc_with_json 00:05:43.608 ************************************ 00:05:43.609 14:04:48 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:05:43.609 14:04:48 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:43.609 14:04:48 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:43.609 14:04:48 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:43.609 ************************************ 00:05:43.609 START TEST skip_rpc_with_delay 00:05:43.609 ************************************ 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:43.609 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:43.609 [2024-09-30 14:04:48.207443] app.c: 840:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:05:43.609 [2024-09-30 14:04:48.207622] app.c: 719:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:43.869 00:05:43.869 real 0m0.178s 00:05:43.869 user 0m0.082s 00:05:43.869 sys 0m0.095s 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:43.869 14:04:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:05:43.869 ************************************ 00:05:43.869 END TEST skip_rpc_with_delay 00:05:43.869 ************************************ 00:05:43.869 14:04:48 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:05:43.869 14:04:48 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:05:43.869 14:04:48 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:05:43.869 14:04:48 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:43.869 14:04:48 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:43.869 14:04:48 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:43.869 ************************************ 00:05:43.869 START TEST exit_on_failed_rpc_init 00:05:43.869 ************************************ 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=70607 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 70607 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 70607 ']' 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:43.869 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:43.869 14:04:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:43.869 [2024-09-30 14:04:48.443232] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:43.869 [2024-09-30 14:04:48.443359] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70607 ] 00:05:44.129 [2024-09-30 14:04:48.573283] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:44.129 [2024-09-30 14:04:48.599439] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:44.129 [2024-09-30 14:04:48.645455] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:44.697 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:44.697 [2024-09-30 14:04:49.343174] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:44.697 [2024-09-30 14:04:49.343342] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70622 ] 00:05:44.955 [2024-09-30 14:04:49.497367] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:44.955 [2024-09-30 14:04:49.525363] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:44.955 [2024-09-30 14:04:49.599859] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:44.955 [2024-09-30 14:04:49.599981] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:05:44.955 [2024-09-30 14:04:49.599998] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:05:44.955 [2024-09-30 14:04:49.600011] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:45.213 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:05:45.213 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 70607 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 70607 ']' 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 70607 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70607 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:45.214 killing process with pid 70607 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70607' 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 70607 00:05:45.214 14:04:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 70607 00:05:45.782 00:05:45.782 real 0m1.839s 00:05:45.782 user 0m1.998s 00:05:45.782 sys 0m0.564s 00:05:45.782 14:04:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:45.782 14:04:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:45.782 ************************************ 00:05:45.782 END TEST exit_on_failed_rpc_init 00:05:45.782 ************************************ 00:05:45.782 14:04:50 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:45.782 00:05:45.782 real 0m14.931s 00:05:45.782 user 0m13.753s 00:05:45.782 sys 0m2.079s 00:05:45.782 14:04:50 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:45.782 14:04:50 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:45.782 ************************************ 00:05:45.782 END TEST skip_rpc 00:05:45.782 ************************************ 00:05:45.782 14:04:50 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:45.782 14:04:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:45.782 14:04:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:45.782 14:04:50 -- common/autotest_common.sh@10 -- # set +x 00:05:45.782 ************************************ 00:05:45.782 START TEST rpc_client 00:05:45.782 ************************************ 00:05:45.783 14:04:50 rpc_client -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:45.783 * Looking for test storage... 00:05:45.783 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:05:45.783 14:04:50 rpc_client -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:45.783 14:04:50 rpc_client -- common/autotest_common.sh@1681 -- # lcov --version 00:05:45.783 14:04:50 rpc_client -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@345 -- # : 1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@353 -- # local d=1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@355 -- # echo 1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@353 -- # local d=2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@355 -- # echo 2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.043 14:04:50 rpc_client -- scripts/common.sh@368 -- # return 0 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:46.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.043 --rc genhtml_branch_coverage=1 00:05:46.043 --rc genhtml_function_coverage=1 00:05:46.043 --rc genhtml_legend=1 00:05:46.043 --rc geninfo_all_blocks=1 00:05:46.043 --rc geninfo_unexecuted_blocks=1 00:05:46.043 00:05:46.043 ' 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:46.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.043 --rc genhtml_branch_coverage=1 00:05:46.043 --rc genhtml_function_coverage=1 00:05:46.043 --rc genhtml_legend=1 00:05:46.043 --rc geninfo_all_blocks=1 00:05:46.043 --rc geninfo_unexecuted_blocks=1 00:05:46.043 00:05:46.043 ' 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:46.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.043 --rc genhtml_branch_coverage=1 00:05:46.043 --rc genhtml_function_coverage=1 00:05:46.043 --rc genhtml_legend=1 00:05:46.043 --rc geninfo_all_blocks=1 00:05:46.043 --rc geninfo_unexecuted_blocks=1 00:05:46.043 00:05:46.043 ' 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:46.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.043 --rc genhtml_branch_coverage=1 00:05:46.043 --rc genhtml_function_coverage=1 00:05:46.043 --rc genhtml_legend=1 00:05:46.043 --rc geninfo_all_blocks=1 00:05:46.043 --rc geninfo_unexecuted_blocks=1 00:05:46.043 00:05:46.043 ' 00:05:46.043 14:04:50 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:05:46.043 OK 00:05:46.043 14:04:50 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:05:46.043 00:05:46.043 real 0m0.287s 00:05:46.043 user 0m0.156s 00:05:46.043 sys 0m0.149s 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.043 14:04:50 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:05:46.043 ************************************ 00:05:46.043 END TEST rpc_client 00:05:46.043 ************************************ 00:05:46.043 14:04:50 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:46.043 14:04:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:46.043 14:04:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:46.043 14:04:50 -- common/autotest_common.sh@10 -- # set +x 00:05:46.043 ************************************ 00:05:46.043 START TEST json_config 00:05:46.043 ************************************ 00:05:46.043 14:04:50 json_config -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1681 -- # lcov --version 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.304 14:04:50 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.304 14:04:50 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.304 14:04:50 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.304 14:04:50 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.304 14:04:50 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.304 14:04:50 json_config -- scripts/common.sh@344 -- # case "$op" in 00:05:46.304 14:04:50 json_config -- scripts/common.sh@345 -- # : 1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.304 14:04:50 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.304 14:04:50 json_config -- scripts/common.sh@365 -- # decimal 1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@353 -- # local d=1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.304 14:04:50 json_config -- scripts/common.sh@355 -- # echo 1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.304 14:04:50 json_config -- scripts/common.sh@366 -- # decimal 2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@353 -- # local d=2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.304 14:04:50 json_config -- scripts/common.sh@355 -- # echo 2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.304 14:04:50 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.304 14:04:50 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.304 14:04:50 json_config -- scripts/common.sh@368 -- # return 0 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:46.304 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.304 --rc genhtml_branch_coverage=1 00:05:46.304 --rc genhtml_function_coverage=1 00:05:46.304 --rc genhtml_legend=1 00:05:46.304 --rc geninfo_all_blocks=1 00:05:46.304 --rc geninfo_unexecuted_blocks=1 00:05:46.304 00:05:46.304 ' 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:46.304 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.304 --rc genhtml_branch_coverage=1 00:05:46.304 --rc genhtml_function_coverage=1 00:05:46.304 --rc genhtml_legend=1 00:05:46.304 --rc geninfo_all_blocks=1 00:05:46.304 --rc geninfo_unexecuted_blocks=1 00:05:46.304 00:05:46.304 ' 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:46.304 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.304 --rc genhtml_branch_coverage=1 00:05:46.304 --rc genhtml_function_coverage=1 00:05:46.304 --rc genhtml_legend=1 00:05:46.304 --rc geninfo_all_blocks=1 00:05:46.304 --rc geninfo_unexecuted_blocks=1 00:05:46.304 00:05:46.304 ' 00:05:46.304 14:04:50 json_config -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:46.304 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.304 --rc genhtml_branch_coverage=1 00:05:46.304 --rc genhtml_function_coverage=1 00:05:46.304 --rc genhtml_legend=1 00:05:46.304 --rc geninfo_all_blocks=1 00:05:46.304 --rc geninfo_unexecuted_blocks=1 00:05:46.304 00:05:46.304 ' 00:05:46.304 14:04:50 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@7 -- # uname -s 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:a1e24555-614e-41f5-8ac1-aaa73d42684b 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=a1e24555-614e-41f5-8ac1-aaa73d42684b 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:46.304 14:04:50 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:46.304 14:04:50 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:05:46.304 14:04:50 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:46.304 14:04:50 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:46.304 14:04:50 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:46.304 14:04:50 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.304 14:04:50 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.304 14:04:50 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.304 14:04:50 json_config -- paths/export.sh@5 -- # export PATH 00:05:46.305 14:04:50 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@51 -- # : 0 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:46.305 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:46.305 14:04:50 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:05:46.305 WARNING: No tests are enabled so not running JSON configuration tests 00:05:46.305 14:04:50 json_config -- json_config/json_config.sh@28 -- # exit 0 00:05:46.305 00:05:46.305 real 0m0.226s 00:05:46.305 user 0m0.135s 00:05:46.305 sys 0m0.095s 00:05:46.305 14:04:50 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.305 14:04:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:05:46.305 ************************************ 00:05:46.305 END TEST json_config 00:05:46.305 ************************************ 00:05:46.305 14:04:50 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:46.305 14:04:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:46.305 14:04:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:46.305 14:04:50 -- common/autotest_common.sh@10 -- # set +x 00:05:46.305 ************************************ 00:05:46.305 START TEST json_config_extra_key 00:05:46.305 ************************************ 00:05:46.305 14:04:50 json_config_extra_key -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1681 -- # lcov --version 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.565 14:04:51 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:46.565 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.565 --rc genhtml_branch_coverage=1 00:05:46.565 --rc genhtml_function_coverage=1 00:05:46.565 --rc genhtml_legend=1 00:05:46.565 --rc geninfo_all_blocks=1 00:05:46.565 --rc geninfo_unexecuted_blocks=1 00:05:46.565 00:05:46.565 ' 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:46.565 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.565 --rc genhtml_branch_coverage=1 00:05:46.565 --rc genhtml_function_coverage=1 00:05:46.565 --rc genhtml_legend=1 00:05:46.565 --rc geninfo_all_blocks=1 00:05:46.565 --rc geninfo_unexecuted_blocks=1 00:05:46.565 00:05:46.565 ' 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:46.565 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.565 --rc genhtml_branch_coverage=1 00:05:46.565 --rc genhtml_function_coverage=1 00:05:46.565 --rc genhtml_legend=1 00:05:46.565 --rc geninfo_all_blocks=1 00:05:46.565 --rc geninfo_unexecuted_blocks=1 00:05:46.565 00:05:46.565 ' 00:05:46.565 14:04:51 json_config_extra_key -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:46.565 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.565 --rc genhtml_branch_coverage=1 00:05:46.565 --rc genhtml_function_coverage=1 00:05:46.565 --rc genhtml_legend=1 00:05:46.565 --rc geninfo_all_blocks=1 00:05:46.565 --rc geninfo_unexecuted_blocks=1 00:05:46.565 00:05:46.565 ' 00:05:46.565 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:46.565 14:04:51 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:05:46.565 14:04:51 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:46.565 14:04:51 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:46.565 14:04:51 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:a1e24555-614e-41f5-8ac1-aaa73d42684b 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=a1e24555-614e-41f5-8ac1-aaa73d42684b 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:46.566 14:04:51 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:05:46.566 14:04:51 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:46.566 14:04:51 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:46.566 14:04:51 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:46.566 14:04:51 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.566 14:04:51 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.566 14:04:51 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.566 14:04:51 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:05:46.566 14:04:51 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:46.566 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:46.566 14:04:51 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:05:46.566 INFO: launching applications... 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:05:46.566 14:04:51 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=70809 00:05:46.566 Waiting for target to run... 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 70809 /var/tmp/spdk_tgt.sock 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 70809 ']' 00:05:46.566 14:04:51 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:46.566 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:46.566 14:04:51 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:46.826 [2024-09-30 14:04:51.269404] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:46.826 [2024-09-30 14:04:51.269531] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70809 ] 00:05:47.086 [2024-09-30 14:04:51.612151] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:47.086 [2024-09-30 14:04:51.641808] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:47.086 [2024-09-30 14:04:51.672507] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:47.655 14:04:52 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:47.655 14:04:52 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:05:47.655 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:05:47.655 INFO: shutting down applications... 00:05:47.655 14:04:52 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:05:47.655 14:04:52 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 70809 ]] 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 70809 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 70809 00:05:47.655 14:04:52 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:05:47.916 14:04:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:05:47.916 14:04:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:47.916 14:04:52 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 70809 00:05:47.916 14:04:52 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:05:48.175 14:04:52 json_config_extra_key -- json_config/common.sh@43 -- # break 00:05:48.175 14:04:52 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:05:48.175 SPDK target shutdown done 00:05:48.175 14:04:52 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:05:48.175 Success 00:05:48.175 14:04:52 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:05:48.175 00:05:48.175 real 0m1.619s 00:05:48.175 user 0m1.305s 00:05:48.175 sys 0m0.492s 00:05:48.175 14:04:52 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.175 14:04:52 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:48.175 ************************************ 00:05:48.175 END TEST json_config_extra_key 00:05:48.175 ************************************ 00:05:48.175 14:04:52 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:48.175 14:04:52 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:48.175 14:04:52 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:48.175 14:04:52 -- common/autotest_common.sh@10 -- # set +x 00:05:48.175 ************************************ 00:05:48.175 START TEST alias_rpc 00:05:48.175 ************************************ 00:05:48.175 14:04:52 alias_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:48.175 * Looking for test storage... 00:05:48.175 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:05:48.175 14:04:52 alias_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:48.175 14:04:52 alias_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:48.175 14:04:52 alias_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:48.435 14:04:52 alias_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@345 -- # : 1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:48.435 14:04:52 alias_rpc -- scripts/common.sh@368 -- # return 0 00:05:48.435 14:04:52 alias_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:48.435 14:04:52 alias_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:48.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.435 --rc genhtml_branch_coverage=1 00:05:48.435 --rc genhtml_function_coverage=1 00:05:48.435 --rc genhtml_legend=1 00:05:48.435 --rc geninfo_all_blocks=1 00:05:48.435 --rc geninfo_unexecuted_blocks=1 00:05:48.435 00:05:48.435 ' 00:05:48.435 14:04:52 alias_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:48.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.436 --rc genhtml_branch_coverage=1 00:05:48.436 --rc genhtml_function_coverage=1 00:05:48.436 --rc genhtml_legend=1 00:05:48.436 --rc geninfo_all_blocks=1 00:05:48.436 --rc geninfo_unexecuted_blocks=1 00:05:48.436 00:05:48.436 ' 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:48.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.436 --rc genhtml_branch_coverage=1 00:05:48.436 --rc genhtml_function_coverage=1 00:05:48.436 --rc genhtml_legend=1 00:05:48.436 --rc geninfo_all_blocks=1 00:05:48.436 --rc geninfo_unexecuted_blocks=1 00:05:48.436 00:05:48.436 ' 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:48.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.436 --rc genhtml_branch_coverage=1 00:05:48.436 --rc genhtml_function_coverage=1 00:05:48.436 --rc genhtml_legend=1 00:05:48.436 --rc geninfo_all_blocks=1 00:05:48.436 --rc geninfo_unexecuted_blocks=1 00:05:48.436 00:05:48.436 ' 00:05:48.436 14:04:52 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:05:48.436 14:04:52 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=70884 00:05:48.436 14:04:52 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:48.436 14:04:52 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 70884 00:05:48.436 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 70884 ']' 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:48.436 14:04:52 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.436 [2024-09-30 14:04:52.966140] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:48.436 [2024-09-30 14:04:52.966357] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70884 ] 00:05:48.696 [2024-09-30 14:04:53.102544] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:48.696 [2024-09-30 14:04:53.129164] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:48.696 [2024-09-30 14:04:53.173166] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.265 14:04:53 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:49.265 14:04:53 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:05:49.265 14:04:53 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:05:49.525 14:04:53 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 70884 00:05:49.525 14:04:53 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 70884 ']' 00:05:49.525 14:04:53 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 70884 00:05:49.525 14:04:53 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:05:49.525 14:04:53 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:49.525 14:04:53 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70884 00:05:49.525 14:04:54 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:49.525 killing process with pid 70884 00:05:49.525 14:04:54 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:49.525 14:04:54 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70884' 00:05:49.525 14:04:54 alias_rpc -- common/autotest_common.sh@969 -- # kill 70884 00:05:49.525 14:04:54 alias_rpc -- common/autotest_common.sh@974 -- # wait 70884 00:05:49.786 ************************************ 00:05:49.786 END TEST alias_rpc 00:05:49.786 ************************************ 00:05:49.786 00:05:49.786 real 0m1.780s 00:05:49.786 user 0m1.771s 00:05:49.786 sys 0m0.537s 00:05:49.786 14:04:54 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:49.786 14:04:54 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:50.046 14:04:54 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:05:50.046 14:04:54 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:50.046 14:04:54 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:50.046 14:04:54 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:50.046 14:04:54 -- common/autotest_common.sh@10 -- # set +x 00:05:50.046 ************************************ 00:05:50.046 START TEST spdkcli_tcp 00:05:50.046 ************************************ 00:05:50.046 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:50.046 * Looking for test storage... 00:05:50.046 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:05:50.046 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:50.046 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lcov --version 00:05:50.046 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:50.046 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:50.046 14:04:54 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:50.307 14:04:54 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:05:50.307 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:50.307 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:50.307 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.307 --rc genhtml_branch_coverage=1 00:05:50.307 --rc genhtml_function_coverage=1 00:05:50.307 --rc genhtml_legend=1 00:05:50.307 --rc geninfo_all_blocks=1 00:05:50.307 --rc geninfo_unexecuted_blocks=1 00:05:50.307 00:05:50.307 ' 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:50.308 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.308 --rc genhtml_branch_coverage=1 00:05:50.308 --rc genhtml_function_coverage=1 00:05:50.308 --rc genhtml_legend=1 00:05:50.308 --rc geninfo_all_blocks=1 00:05:50.308 --rc geninfo_unexecuted_blocks=1 00:05:50.308 00:05:50.308 ' 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:50.308 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.308 --rc genhtml_branch_coverage=1 00:05:50.308 --rc genhtml_function_coverage=1 00:05:50.308 --rc genhtml_legend=1 00:05:50.308 --rc geninfo_all_blocks=1 00:05:50.308 --rc geninfo_unexecuted_blocks=1 00:05:50.308 00:05:50.308 ' 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:50.308 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.308 --rc genhtml_branch_coverage=1 00:05:50.308 --rc genhtml_function_coverage=1 00:05:50.308 --rc genhtml_legend=1 00:05:50.308 --rc geninfo_all_blocks=1 00:05:50.308 --rc geninfo_unexecuted_blocks=1 00:05:50.308 00:05:50.308 ' 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=70969 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:05:50.308 14:04:54 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 70969 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 70969 ']' 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:50.308 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:50.308 14:04:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:50.308 [2024-09-30 14:04:54.820326] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:50.308 [2024-09-30 14:04:54.820451] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70969 ] 00:05:50.308 [2024-09-30 14:04:54.952874] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:50.567 [2024-09-30 14:04:54.980957] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:50.567 [2024-09-30 14:04:55.027295] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:50.567 [2024-09-30 14:04:55.027373] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:51.136 14:04:55 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:51.136 14:04:55 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:05:51.136 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:05:51.136 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=70986 00:05:51.136 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:05:51.396 [ 00:05:51.397 "bdev_malloc_delete", 00:05:51.397 "bdev_malloc_create", 00:05:51.397 "bdev_null_resize", 00:05:51.397 "bdev_null_delete", 00:05:51.397 "bdev_null_create", 00:05:51.397 "bdev_nvme_cuse_unregister", 00:05:51.397 "bdev_nvme_cuse_register", 00:05:51.397 "bdev_opal_new_user", 00:05:51.397 "bdev_opal_set_lock_state", 00:05:51.397 "bdev_opal_delete", 00:05:51.397 "bdev_opal_get_info", 00:05:51.397 "bdev_opal_create", 00:05:51.397 "bdev_nvme_opal_revert", 00:05:51.397 "bdev_nvme_opal_init", 00:05:51.397 "bdev_nvme_send_cmd", 00:05:51.397 "bdev_nvme_set_keys", 00:05:51.397 "bdev_nvme_get_path_iostat", 00:05:51.397 "bdev_nvme_get_mdns_discovery_info", 00:05:51.397 "bdev_nvme_stop_mdns_discovery", 00:05:51.397 "bdev_nvme_start_mdns_discovery", 00:05:51.397 "bdev_nvme_set_multipath_policy", 00:05:51.397 "bdev_nvme_set_preferred_path", 00:05:51.397 "bdev_nvme_get_io_paths", 00:05:51.397 "bdev_nvme_remove_error_injection", 00:05:51.397 "bdev_nvme_add_error_injection", 00:05:51.397 "bdev_nvme_get_discovery_info", 00:05:51.397 "bdev_nvme_stop_discovery", 00:05:51.397 "bdev_nvme_start_discovery", 00:05:51.397 "bdev_nvme_get_controller_health_info", 00:05:51.397 "bdev_nvme_disable_controller", 00:05:51.397 "bdev_nvme_enable_controller", 00:05:51.397 "bdev_nvme_reset_controller", 00:05:51.397 "bdev_nvme_get_transport_statistics", 00:05:51.397 "bdev_nvme_apply_firmware", 00:05:51.397 "bdev_nvme_detach_controller", 00:05:51.397 "bdev_nvme_get_controllers", 00:05:51.397 "bdev_nvme_attach_controller", 00:05:51.397 "bdev_nvme_set_hotplug", 00:05:51.397 "bdev_nvme_set_options", 00:05:51.397 "bdev_passthru_delete", 00:05:51.397 "bdev_passthru_create", 00:05:51.397 "bdev_lvol_set_parent_bdev", 00:05:51.397 "bdev_lvol_set_parent", 00:05:51.397 "bdev_lvol_check_shallow_copy", 00:05:51.397 "bdev_lvol_start_shallow_copy", 00:05:51.397 "bdev_lvol_grow_lvstore", 00:05:51.397 "bdev_lvol_get_lvols", 00:05:51.397 "bdev_lvol_get_lvstores", 00:05:51.397 "bdev_lvol_delete", 00:05:51.397 "bdev_lvol_set_read_only", 00:05:51.397 "bdev_lvol_resize", 00:05:51.397 "bdev_lvol_decouple_parent", 00:05:51.397 "bdev_lvol_inflate", 00:05:51.397 "bdev_lvol_rename", 00:05:51.397 "bdev_lvol_clone_bdev", 00:05:51.397 "bdev_lvol_clone", 00:05:51.397 "bdev_lvol_snapshot", 00:05:51.397 "bdev_lvol_create", 00:05:51.397 "bdev_lvol_delete_lvstore", 00:05:51.397 "bdev_lvol_rename_lvstore", 00:05:51.397 "bdev_lvol_create_lvstore", 00:05:51.397 "bdev_raid_set_options", 00:05:51.397 "bdev_raid_remove_base_bdev", 00:05:51.397 "bdev_raid_add_base_bdev", 00:05:51.397 "bdev_raid_delete", 00:05:51.397 "bdev_raid_create", 00:05:51.397 "bdev_raid_get_bdevs", 00:05:51.397 "bdev_error_inject_error", 00:05:51.397 "bdev_error_delete", 00:05:51.397 "bdev_error_create", 00:05:51.397 "bdev_split_delete", 00:05:51.397 "bdev_split_create", 00:05:51.397 "bdev_delay_delete", 00:05:51.397 "bdev_delay_create", 00:05:51.397 "bdev_delay_update_latency", 00:05:51.397 "bdev_zone_block_delete", 00:05:51.397 "bdev_zone_block_create", 00:05:51.397 "blobfs_create", 00:05:51.397 "blobfs_detect", 00:05:51.397 "blobfs_set_cache_size", 00:05:51.397 "bdev_aio_delete", 00:05:51.397 "bdev_aio_rescan", 00:05:51.397 "bdev_aio_create", 00:05:51.397 "bdev_ftl_set_property", 00:05:51.397 "bdev_ftl_get_properties", 00:05:51.397 "bdev_ftl_get_stats", 00:05:51.397 "bdev_ftl_unmap", 00:05:51.397 "bdev_ftl_unload", 00:05:51.397 "bdev_ftl_delete", 00:05:51.397 "bdev_ftl_load", 00:05:51.397 "bdev_ftl_create", 00:05:51.397 "bdev_virtio_attach_controller", 00:05:51.397 "bdev_virtio_scsi_get_devices", 00:05:51.397 "bdev_virtio_detach_controller", 00:05:51.397 "bdev_virtio_blk_set_hotplug", 00:05:51.397 "bdev_iscsi_delete", 00:05:51.397 "bdev_iscsi_create", 00:05:51.397 "bdev_iscsi_set_options", 00:05:51.397 "accel_error_inject_error", 00:05:51.397 "ioat_scan_accel_module", 00:05:51.397 "dsa_scan_accel_module", 00:05:51.397 "iaa_scan_accel_module", 00:05:51.397 "keyring_file_remove_key", 00:05:51.397 "keyring_file_add_key", 00:05:51.397 "keyring_linux_set_options", 00:05:51.397 "fsdev_aio_delete", 00:05:51.397 "fsdev_aio_create", 00:05:51.397 "iscsi_get_histogram", 00:05:51.397 "iscsi_enable_histogram", 00:05:51.397 "iscsi_set_options", 00:05:51.397 "iscsi_get_auth_groups", 00:05:51.397 "iscsi_auth_group_remove_secret", 00:05:51.397 "iscsi_auth_group_add_secret", 00:05:51.397 "iscsi_delete_auth_group", 00:05:51.397 "iscsi_create_auth_group", 00:05:51.397 "iscsi_set_discovery_auth", 00:05:51.397 "iscsi_get_options", 00:05:51.397 "iscsi_target_node_request_logout", 00:05:51.397 "iscsi_target_node_set_redirect", 00:05:51.397 "iscsi_target_node_set_auth", 00:05:51.397 "iscsi_target_node_add_lun", 00:05:51.397 "iscsi_get_stats", 00:05:51.397 "iscsi_get_connections", 00:05:51.397 "iscsi_portal_group_set_auth", 00:05:51.397 "iscsi_start_portal_group", 00:05:51.397 "iscsi_delete_portal_group", 00:05:51.397 "iscsi_create_portal_group", 00:05:51.397 "iscsi_get_portal_groups", 00:05:51.397 "iscsi_delete_target_node", 00:05:51.397 "iscsi_target_node_remove_pg_ig_maps", 00:05:51.397 "iscsi_target_node_add_pg_ig_maps", 00:05:51.397 "iscsi_create_target_node", 00:05:51.397 "iscsi_get_target_nodes", 00:05:51.397 "iscsi_delete_initiator_group", 00:05:51.397 "iscsi_initiator_group_remove_initiators", 00:05:51.397 "iscsi_initiator_group_add_initiators", 00:05:51.397 "iscsi_create_initiator_group", 00:05:51.397 "iscsi_get_initiator_groups", 00:05:51.397 "nvmf_set_crdt", 00:05:51.397 "nvmf_set_config", 00:05:51.397 "nvmf_set_max_subsystems", 00:05:51.397 "nvmf_stop_mdns_prr", 00:05:51.397 "nvmf_publish_mdns_prr", 00:05:51.397 "nvmf_subsystem_get_listeners", 00:05:51.397 "nvmf_subsystem_get_qpairs", 00:05:51.397 "nvmf_subsystem_get_controllers", 00:05:51.397 "nvmf_get_stats", 00:05:51.397 "nvmf_get_transports", 00:05:51.397 "nvmf_create_transport", 00:05:51.397 "nvmf_get_targets", 00:05:51.397 "nvmf_delete_target", 00:05:51.397 "nvmf_create_target", 00:05:51.397 "nvmf_subsystem_allow_any_host", 00:05:51.397 "nvmf_subsystem_set_keys", 00:05:51.397 "nvmf_subsystem_remove_host", 00:05:51.397 "nvmf_subsystem_add_host", 00:05:51.397 "nvmf_ns_remove_host", 00:05:51.397 "nvmf_ns_add_host", 00:05:51.397 "nvmf_subsystem_remove_ns", 00:05:51.397 "nvmf_subsystem_set_ns_ana_group", 00:05:51.397 "nvmf_subsystem_add_ns", 00:05:51.397 "nvmf_subsystem_listener_set_ana_state", 00:05:51.397 "nvmf_discovery_get_referrals", 00:05:51.397 "nvmf_discovery_remove_referral", 00:05:51.397 "nvmf_discovery_add_referral", 00:05:51.397 "nvmf_subsystem_remove_listener", 00:05:51.397 "nvmf_subsystem_add_listener", 00:05:51.397 "nvmf_delete_subsystem", 00:05:51.397 "nvmf_create_subsystem", 00:05:51.397 "nvmf_get_subsystems", 00:05:51.397 "env_dpdk_get_mem_stats", 00:05:51.397 "nbd_get_disks", 00:05:51.397 "nbd_stop_disk", 00:05:51.397 "nbd_start_disk", 00:05:51.397 "ublk_recover_disk", 00:05:51.397 "ublk_get_disks", 00:05:51.397 "ublk_stop_disk", 00:05:51.397 "ublk_start_disk", 00:05:51.397 "ublk_destroy_target", 00:05:51.397 "ublk_create_target", 00:05:51.397 "virtio_blk_create_transport", 00:05:51.397 "virtio_blk_get_transports", 00:05:51.397 "vhost_controller_set_coalescing", 00:05:51.397 "vhost_get_controllers", 00:05:51.397 "vhost_delete_controller", 00:05:51.397 "vhost_create_blk_controller", 00:05:51.397 "vhost_scsi_controller_remove_target", 00:05:51.397 "vhost_scsi_controller_add_target", 00:05:51.397 "vhost_start_scsi_controller", 00:05:51.397 "vhost_create_scsi_controller", 00:05:51.397 "thread_set_cpumask", 00:05:51.397 "scheduler_set_options", 00:05:51.397 "framework_get_governor", 00:05:51.397 "framework_get_scheduler", 00:05:51.397 "framework_set_scheduler", 00:05:51.397 "framework_get_reactors", 00:05:51.397 "thread_get_io_channels", 00:05:51.397 "thread_get_pollers", 00:05:51.397 "thread_get_stats", 00:05:51.397 "framework_monitor_context_switch", 00:05:51.397 "spdk_kill_instance", 00:05:51.397 "log_enable_timestamps", 00:05:51.397 "log_get_flags", 00:05:51.397 "log_clear_flag", 00:05:51.397 "log_set_flag", 00:05:51.397 "log_get_level", 00:05:51.397 "log_set_level", 00:05:51.397 "log_get_print_level", 00:05:51.397 "log_set_print_level", 00:05:51.397 "framework_enable_cpumask_locks", 00:05:51.397 "framework_disable_cpumask_locks", 00:05:51.397 "framework_wait_init", 00:05:51.397 "framework_start_init", 00:05:51.397 "scsi_get_devices", 00:05:51.397 "bdev_get_histogram", 00:05:51.397 "bdev_enable_histogram", 00:05:51.397 "bdev_set_qos_limit", 00:05:51.397 "bdev_set_qd_sampling_period", 00:05:51.397 "bdev_get_bdevs", 00:05:51.397 "bdev_reset_iostat", 00:05:51.397 "bdev_get_iostat", 00:05:51.397 "bdev_examine", 00:05:51.397 "bdev_wait_for_examine", 00:05:51.397 "bdev_set_options", 00:05:51.397 "accel_get_stats", 00:05:51.397 "accel_set_options", 00:05:51.397 "accel_set_driver", 00:05:51.397 "accel_crypto_key_destroy", 00:05:51.397 "accel_crypto_keys_get", 00:05:51.397 "accel_crypto_key_create", 00:05:51.398 "accel_assign_opc", 00:05:51.398 "accel_get_module_info", 00:05:51.398 "accel_get_opc_assignments", 00:05:51.398 "vmd_rescan", 00:05:51.398 "vmd_remove_device", 00:05:51.398 "vmd_enable", 00:05:51.398 "sock_get_default_impl", 00:05:51.398 "sock_set_default_impl", 00:05:51.398 "sock_impl_set_options", 00:05:51.398 "sock_impl_get_options", 00:05:51.398 "iobuf_get_stats", 00:05:51.398 "iobuf_set_options", 00:05:51.398 "keyring_get_keys", 00:05:51.398 "framework_get_pci_devices", 00:05:51.398 "framework_get_config", 00:05:51.398 "framework_get_subsystems", 00:05:51.398 "fsdev_set_opts", 00:05:51.398 "fsdev_get_opts", 00:05:51.398 "trace_get_info", 00:05:51.398 "trace_get_tpoint_group_mask", 00:05:51.398 "trace_disable_tpoint_group", 00:05:51.398 "trace_enable_tpoint_group", 00:05:51.398 "trace_clear_tpoint_mask", 00:05:51.398 "trace_set_tpoint_mask", 00:05:51.398 "notify_get_notifications", 00:05:51.398 "notify_get_types", 00:05:51.398 "spdk_get_version", 00:05:51.398 "rpc_get_methods" 00:05:51.398 ] 00:05:51.398 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:51.398 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:05:51.398 14:04:55 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 70969 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 70969 ']' 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 70969 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70969 00:05:51.398 killing process with pid 70969 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70969' 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 70969 00:05:51.398 14:04:55 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 70969 00:05:51.658 ************************************ 00:05:51.659 END TEST spdkcli_tcp 00:05:51.659 ************************************ 00:05:51.659 00:05:51.659 real 0m1.817s 00:05:51.659 user 0m2.945s 00:05:51.659 sys 0m0.587s 00:05:51.659 14:04:56 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:51.659 14:04:56 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:51.920 14:04:56 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:51.920 14:04:56 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:51.920 14:04:56 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:51.920 14:04:56 -- common/autotest_common.sh@10 -- # set +x 00:05:51.920 ************************************ 00:05:51.920 START TEST dpdk_mem_utility 00:05:51.920 ************************************ 00:05:51.920 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:51.920 * Looking for test storage... 00:05:51.920 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:05:51.920 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:51.920 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lcov --version 00:05:51.920 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:52.181 14:04:56 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:52.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.181 --rc genhtml_branch_coverage=1 00:05:52.181 --rc genhtml_function_coverage=1 00:05:52.181 --rc genhtml_legend=1 00:05:52.181 --rc geninfo_all_blocks=1 00:05:52.181 --rc geninfo_unexecuted_blocks=1 00:05:52.181 00:05:52.181 ' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:52.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.181 --rc genhtml_branch_coverage=1 00:05:52.181 --rc genhtml_function_coverage=1 00:05:52.181 --rc genhtml_legend=1 00:05:52.181 --rc geninfo_all_blocks=1 00:05:52.181 --rc geninfo_unexecuted_blocks=1 00:05:52.181 00:05:52.181 ' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:52.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.181 --rc genhtml_branch_coverage=1 00:05:52.181 --rc genhtml_function_coverage=1 00:05:52.181 --rc genhtml_legend=1 00:05:52.181 --rc geninfo_all_blocks=1 00:05:52.181 --rc geninfo_unexecuted_blocks=1 00:05:52.181 00:05:52.181 ' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:52.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.181 --rc genhtml_branch_coverage=1 00:05:52.181 --rc genhtml_function_coverage=1 00:05:52.181 --rc genhtml_legend=1 00:05:52.181 --rc geninfo_all_blocks=1 00:05:52.181 --rc geninfo_unexecuted_blocks=1 00:05:52.181 00:05:52.181 ' 00:05:52.181 14:04:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:52.181 14:04:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=71068 00:05:52.181 14:04:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:52.181 14:04:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 71068 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 71068 ']' 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:52.181 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:52.181 14:04:56 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:52.181 [2024-09-30 14:04:56.684110] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:52.181 [2024-09-30 14:04:56.684325] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71068 ] 00:05:52.181 [2024-09-30 14:04:56.814903] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:52.441 [2024-09-30 14:04:56.842329] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:52.441 [2024-09-30 14:04:56.886140] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:53.012 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:53.012 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:05:53.012 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:05:53.012 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:05:53.012 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:53.012 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:53.012 { 00:05:53.012 "filename": "/tmp/spdk_mem_dump.txt" 00:05:53.012 } 00:05:53.012 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:53.012 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:53.012 DPDK memory size 860.000000 MiB in 1 heap(s) 00:05:53.012 1 heaps totaling size 860.000000 MiB 00:05:53.012 size: 860.000000 MiB heap id: 0 00:05:53.012 end heaps---------- 00:05:53.012 9 mempools totaling size 642.649841 MiB 00:05:53.012 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:05:53.012 size: 158.602051 MiB name: PDU_data_out_Pool 00:05:53.012 size: 92.545471 MiB name: bdev_io_71068 00:05:53.012 size: 51.011292 MiB name: evtpool_71068 00:05:53.012 size: 50.003479 MiB name: msgpool_71068 00:05:53.012 size: 36.509338 MiB name: fsdev_io_71068 00:05:53.012 size: 21.763794 MiB name: PDU_Pool 00:05:53.012 size: 19.513306 MiB name: SCSI_TASK_Pool 00:05:53.012 size: 0.026123 MiB name: Session_Pool 00:05:53.012 end mempools------- 00:05:53.012 6 memzones totaling size 4.142822 MiB 00:05:53.012 size: 1.000366 MiB name: RG_ring_0_71068 00:05:53.012 size: 1.000366 MiB name: RG_ring_1_71068 00:05:53.012 size: 1.000366 MiB name: RG_ring_4_71068 00:05:53.012 size: 1.000366 MiB name: RG_ring_5_71068 00:05:53.012 size: 0.125366 MiB name: RG_ring_2_71068 00:05:53.012 size: 0.015991 MiB name: RG_ring_3_71068 00:05:53.012 end memzones------- 00:05:53.012 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:05:53.012 heap id: 0 total size: 860.000000 MiB number of busy elements: 311 number of free elements: 16 00:05:53.012 list of free elements. size: 13.935791 MiB 00:05:53.012 element at address: 0x200000400000 with size: 1.999512 MiB 00:05:53.012 element at address: 0x200000800000 with size: 1.996948 MiB 00:05:53.012 element at address: 0x20001bc00000 with size: 0.999878 MiB 00:05:53.012 element at address: 0x20001be00000 with size: 0.999878 MiB 00:05:53.012 element at address: 0x200034a00000 with size: 0.994446 MiB 00:05:53.012 element at address: 0x200009600000 with size: 0.959839 MiB 00:05:53.012 element at address: 0x200015e00000 with size: 0.954285 MiB 00:05:53.012 element at address: 0x20001c000000 with size: 0.936584 MiB 00:05:53.012 element at address: 0x200000200000 with size: 0.834839 MiB 00:05:53.012 element at address: 0x20001d800000 with size: 0.568237 MiB 00:05:53.012 element at address: 0x20000d800000 with size: 0.489258 MiB 00:05:53.012 element at address: 0x200003e00000 with size: 0.487183 MiB 00:05:53.012 element at address: 0x20001c200000 with size: 0.485657 MiB 00:05:53.012 element at address: 0x200007000000 with size: 0.480469 MiB 00:05:53.012 element at address: 0x20002ac00000 with size: 0.395752 MiB 00:05:53.012 element at address: 0x200003a00000 with size: 0.353027 MiB 00:05:53.012 list of standard malloc elements. size: 199.267517 MiB 00:05:53.012 element at address: 0x20000d9fff80 with size: 132.000122 MiB 00:05:53.012 element at address: 0x2000097fff80 with size: 64.000122 MiB 00:05:53.012 element at address: 0x20001bcfff80 with size: 1.000122 MiB 00:05:53.012 element at address: 0x20001befff80 with size: 1.000122 MiB 00:05:53.012 element at address: 0x20001c0fff80 with size: 1.000122 MiB 00:05:53.012 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:05:53.012 element at address: 0x20001c0eff00 with size: 0.062622 MiB 00:05:53.012 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:05:53.012 element at address: 0x20001c0efdc0 with size: 0.000305 MiB 00:05:53.012 element at address: 0x2000002d5b80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d5c40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d5d00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d5dc0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d5e80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d5f40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6000 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d60c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6180 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6240 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6300 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d63c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6480 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6540 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6600 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d66c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d68c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6980 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6a40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6b00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6bc0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6c80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6d40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6e00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6ec0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d6f80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7040 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7100 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d71c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7280 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7340 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7400 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d74c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7580 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7640 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7700 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d77c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7880 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7940 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7a00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7ac0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7b80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a5a600 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a5a800 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a5eac0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7ed80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7ee40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7ef00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7efc0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f080 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f140 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f200 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f2c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f380 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f440 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f500 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003a7f5c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003aff880 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003affa80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003affb40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7cb80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7cc40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7cd00 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7cdc0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7ce80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7cf40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d000 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d0c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d180 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d240 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d300 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d3c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d480 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d540 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d600 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d6c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d780 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d840 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d900 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7d9c0 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7da80 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7db40 with size: 0.000183 MiB 00:05:53.012 element at address: 0x200003e7dc00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7dcc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7dd80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7de40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7df00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7dfc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e080 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e140 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e200 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e2c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e380 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e440 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e500 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e5c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e680 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e740 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e800 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e8c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7e980 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7ea40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7eb00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7ebc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7ec80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7ed40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003e7ee00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200003eff0c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b000 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b0c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b180 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b240 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b300 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b3c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b480 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b540 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b600 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000707b6c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x2000070fb980 with size: 0.000183 MiB 00:05:53.013 element at address: 0x2000096fdd80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d400 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d4c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d580 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d640 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d700 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d7c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d880 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87d940 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87da00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d87dac0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20000d8fdd80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x200015ef44c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001c0efc40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001c0efd00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001c2bc740 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891780 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891840 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891900 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8919c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891a80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891b40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891c00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891cc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891d80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891e40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891f00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d891fc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892080 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892140 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892200 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8922c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892380 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892440 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892500 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8925c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892680 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892740 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892800 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8928c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892980 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892a40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892b00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892bc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892c80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892d40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892e00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892ec0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d892f80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893040 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893100 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8931c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893280 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893340 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893400 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8934c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893580 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893640 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893700 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8937c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893880 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893940 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893a00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893ac0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893b80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893c40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893d00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893dc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893e80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d893f40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894000 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8940c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894180 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894240 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894300 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8943c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894480 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894540 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894600 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8946c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894780 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894840 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894900 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8949c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894a80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894b40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894c00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894cc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894d80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894e40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894f00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d894fc0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d895080 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d895140 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d895200 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d8952c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d895380 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20001d895440 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac65500 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac655c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c1c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c3c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c480 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c540 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c600 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c6c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c780 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c840 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c900 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6c9c0 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6ca80 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6cb40 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6cc00 with size: 0.000183 MiB 00:05:53.013 element at address: 0x20002ac6ccc0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6cd80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ce40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6cf00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6cfc0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d080 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d140 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d200 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d2c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d380 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d440 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d500 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d5c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d680 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d740 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d800 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d8c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6d980 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6da40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6db00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6dbc0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6dc80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6dd40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6de00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6dec0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6df80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e040 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e100 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e1c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e280 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e340 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e400 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e4c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e580 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e640 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e700 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e7c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e880 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6e940 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ea00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6eac0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6eb80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ec40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ed00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6edc0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ee80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ef40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f000 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f0c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f180 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f240 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f300 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f3c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f480 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f540 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f600 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f6c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f780 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f840 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f900 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6f9c0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fa80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fb40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fc00 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fcc0 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fd80 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6fe40 with size: 0.000183 MiB 00:05:53.014 element at address: 0x20002ac6ff00 with size: 0.000183 MiB 00:05:53.014 list of memzone associated elements. size: 646.796692 MiB 00:05:53.014 element at address: 0x20001d895500 with size: 211.416748 MiB 00:05:53.014 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:05:53.014 element at address: 0x20002ac6ffc0 with size: 157.562561 MiB 00:05:53.014 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:05:53.014 element at address: 0x200015ff4780 with size: 92.045044 MiB 00:05:53.014 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_71068_0 00:05:53.014 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:05:53.014 associated memzone info: size: 48.002930 MiB name: MP_evtpool_71068_0 00:05:53.014 element at address: 0x200003fff380 with size: 48.003052 MiB 00:05:53.014 associated memzone info: size: 48.002930 MiB name: MP_msgpool_71068_0 00:05:53.014 element at address: 0x2000071fdb80 with size: 36.008911 MiB 00:05:53.014 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_71068_0 00:05:53.014 element at address: 0x20001c3be940 with size: 20.255554 MiB 00:05:53.014 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:05:53.014 element at address: 0x200034bfeb40 with size: 18.005066 MiB 00:05:53.014 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:05:53.014 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:05:53.014 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_71068 00:05:53.014 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:05:53.014 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_71068 00:05:53.014 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:05:53.014 associated memzone info: size: 1.007996 MiB name: MP_evtpool_71068 00:05:53.014 element at address: 0x20000d8fde40 with size: 1.008118 MiB 00:05:53.014 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:05:53.014 element at address: 0x20001c2bc800 with size: 1.008118 MiB 00:05:53.014 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:05:53.014 element at address: 0x2000096fde40 with size: 1.008118 MiB 00:05:53.014 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:05:53.014 element at address: 0x2000070fba40 with size: 1.008118 MiB 00:05:53.014 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:05:53.014 element at address: 0x200003eff180 with size: 1.000488 MiB 00:05:53.014 associated memzone info: size: 1.000366 MiB name: RG_ring_0_71068 00:05:53.014 element at address: 0x200003affc00 with size: 1.000488 MiB 00:05:53.014 associated memzone info: size: 1.000366 MiB name: RG_ring_1_71068 00:05:53.014 element at address: 0x200015ef4580 with size: 1.000488 MiB 00:05:53.014 associated memzone info: size: 1.000366 MiB name: RG_ring_4_71068 00:05:53.014 element at address: 0x200034afe940 with size: 1.000488 MiB 00:05:53.014 associated memzone info: size: 1.000366 MiB name: RG_ring_5_71068 00:05:53.014 element at address: 0x200003a7f680 with size: 0.500488 MiB 00:05:53.014 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_71068 00:05:53.014 element at address: 0x200003e7eec0 with size: 0.500488 MiB 00:05:53.014 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_71068 00:05:53.014 element at address: 0x20000d87db80 with size: 0.500488 MiB 00:05:53.014 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:05:53.014 element at address: 0x20000707b780 with size: 0.500488 MiB 00:05:53.014 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:05:53.014 element at address: 0x20001c27c540 with size: 0.250488 MiB 00:05:53.014 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:05:53.014 element at address: 0x200003a5eb80 with size: 0.125488 MiB 00:05:53.014 associated memzone info: size: 0.125366 MiB name: RG_ring_2_71068 00:05:53.014 element at address: 0x2000096f5b80 with size: 0.031738 MiB 00:05:53.014 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:05:53.014 element at address: 0x20002ac65680 with size: 0.023743 MiB 00:05:53.014 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:05:53.014 element at address: 0x200003a5a8c0 with size: 0.016113 MiB 00:05:53.014 associated memzone info: size: 0.015991 MiB name: RG_ring_3_71068 00:05:53.014 element at address: 0x20002ac6b7c0 with size: 0.002441 MiB 00:05:53.014 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:05:53.014 element at address: 0x2000002d6780 with size: 0.000305 MiB 00:05:53.014 associated memzone info: size: 0.000183 MiB name: MP_msgpool_71068 00:05:53.014 element at address: 0x200003aff940 with size: 0.000305 MiB 00:05:53.014 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_71068 00:05:53.014 element at address: 0x200003a5a6c0 with size: 0.000305 MiB 00:05:53.014 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_71068 00:05:53.014 element at address: 0x20002ac6c280 with size: 0.000305 MiB 00:05:53.014 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:05:53.014 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:05:53.014 14:04:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 71068 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 71068 ']' 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 71068 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71068 00:05:53.014 killing process with pid 71068 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71068' 00:05:53.014 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 71068 00:05:53.015 14:04:57 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 71068 00:05:53.585 00:05:53.585 real 0m1.674s 00:05:53.585 user 0m1.605s 00:05:53.585 sys 0m0.512s 00:05:53.585 14:04:58 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:53.585 14:04:58 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:53.585 ************************************ 00:05:53.585 END TEST dpdk_mem_utility 00:05:53.585 ************************************ 00:05:53.585 14:04:58 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:53.585 14:04:58 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:53.585 14:04:58 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:53.585 14:04:58 -- common/autotest_common.sh@10 -- # set +x 00:05:53.585 ************************************ 00:05:53.585 START TEST event 00:05:53.585 ************************************ 00:05:53.585 14:04:58 event -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:53.585 * Looking for test storage... 00:05:53.585 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:05:53.585 14:04:58 event -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:53.585 14:04:58 event -- common/autotest_common.sh@1681 -- # lcov --version 00:05:53.585 14:04:58 event -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:53.845 14:04:58 event -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:53.845 14:04:58 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:53.845 14:04:58 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:53.846 14:04:58 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:53.846 14:04:58 event -- scripts/common.sh@336 -- # IFS=.-: 00:05:53.846 14:04:58 event -- scripts/common.sh@336 -- # read -ra ver1 00:05:53.846 14:04:58 event -- scripts/common.sh@337 -- # IFS=.-: 00:05:53.846 14:04:58 event -- scripts/common.sh@337 -- # read -ra ver2 00:05:53.846 14:04:58 event -- scripts/common.sh@338 -- # local 'op=<' 00:05:53.846 14:04:58 event -- scripts/common.sh@340 -- # ver1_l=2 00:05:53.846 14:04:58 event -- scripts/common.sh@341 -- # ver2_l=1 00:05:53.846 14:04:58 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:53.846 14:04:58 event -- scripts/common.sh@344 -- # case "$op" in 00:05:53.846 14:04:58 event -- scripts/common.sh@345 -- # : 1 00:05:53.846 14:04:58 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:53.846 14:04:58 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:53.846 14:04:58 event -- scripts/common.sh@365 -- # decimal 1 00:05:53.846 14:04:58 event -- scripts/common.sh@353 -- # local d=1 00:05:53.846 14:04:58 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:53.846 14:04:58 event -- scripts/common.sh@355 -- # echo 1 00:05:53.846 14:04:58 event -- scripts/common.sh@365 -- # ver1[v]=1 00:05:53.846 14:04:58 event -- scripts/common.sh@366 -- # decimal 2 00:05:53.846 14:04:58 event -- scripts/common.sh@353 -- # local d=2 00:05:53.846 14:04:58 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:53.846 14:04:58 event -- scripts/common.sh@355 -- # echo 2 00:05:53.846 14:04:58 event -- scripts/common.sh@366 -- # ver2[v]=2 00:05:53.846 14:04:58 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:53.846 14:04:58 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:53.846 14:04:58 event -- scripts/common.sh@368 -- # return 0 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:53.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:53.846 --rc genhtml_branch_coverage=1 00:05:53.846 --rc genhtml_function_coverage=1 00:05:53.846 --rc genhtml_legend=1 00:05:53.846 --rc geninfo_all_blocks=1 00:05:53.846 --rc geninfo_unexecuted_blocks=1 00:05:53.846 00:05:53.846 ' 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:53.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:53.846 --rc genhtml_branch_coverage=1 00:05:53.846 --rc genhtml_function_coverage=1 00:05:53.846 --rc genhtml_legend=1 00:05:53.846 --rc geninfo_all_blocks=1 00:05:53.846 --rc geninfo_unexecuted_blocks=1 00:05:53.846 00:05:53.846 ' 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:53.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:53.846 --rc genhtml_branch_coverage=1 00:05:53.846 --rc genhtml_function_coverage=1 00:05:53.846 --rc genhtml_legend=1 00:05:53.846 --rc geninfo_all_blocks=1 00:05:53.846 --rc geninfo_unexecuted_blocks=1 00:05:53.846 00:05:53.846 ' 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:53.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:53.846 --rc genhtml_branch_coverage=1 00:05:53.846 --rc genhtml_function_coverage=1 00:05:53.846 --rc genhtml_legend=1 00:05:53.846 --rc geninfo_all_blocks=1 00:05:53.846 --rc geninfo_unexecuted_blocks=1 00:05:53.846 00:05:53.846 ' 00:05:53.846 14:04:58 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:05:53.846 14:04:58 event -- bdev/nbd_common.sh@6 -- # set -e 00:05:53.846 14:04:58 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:05:53.846 14:04:58 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:53.846 14:04:58 event -- common/autotest_common.sh@10 -- # set +x 00:05:53.846 ************************************ 00:05:53.846 START TEST event_perf 00:05:53.846 ************************************ 00:05:53.846 14:04:58 event.event_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:53.846 Running I/O for 1 seconds...[2024-09-30 14:04:58.394905] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:53.846 [2024-09-30 14:04:58.395100] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71144 ] 00:05:54.114 [2024-09-30 14:04:58.529957] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:54.114 [2024-09-30 14:04:58.560091] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:54.114 Running I/O for 1 seconds...[2024-09-30 14:04:58.608355] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:54.114 [2024-09-30 14:04:58.608649] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:54.114 [2024-09-30 14:04:58.608619] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:05:54.114 [2024-09-30 14:04:58.608767] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:05:55.072 00:05:55.072 lcore 0: 95101 00:05:55.072 lcore 1: 95096 00:05:55.072 lcore 2: 95099 00:05:55.072 lcore 3: 95098 00:05:55.072 done. 00:05:55.072 00:05:55.072 real 0m1.360s 00:05:55.072 user 0m4.114s 00:05:55.072 sys 0m0.123s 00:05:55.072 14:04:59 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:55.072 14:04:59 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:05:55.072 ************************************ 00:05:55.072 END TEST event_perf 00:05:55.072 ************************************ 00:05:55.332 14:04:59 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:55.332 14:04:59 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:05:55.332 14:04:59 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:55.332 14:04:59 event -- common/autotest_common.sh@10 -- # set +x 00:05:55.332 ************************************ 00:05:55.332 START TEST event_reactor 00:05:55.332 ************************************ 00:05:55.332 14:04:59 event.event_reactor -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:55.332 [2024-09-30 14:04:59.820985] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:55.332 [2024-09-30 14:04:59.821106] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71189 ] 00:05:55.332 [2024-09-30 14:04:59.955068] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:55.332 [2024-09-30 14:04:59.985187] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:55.591 [2024-09-30 14:05:00.030176] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:56.531 test_start 00:05:56.531 oneshot 00:05:56.531 tick 100 00:05:56.531 tick 100 00:05:56.531 tick 250 00:05:56.531 tick 100 00:05:56.531 tick 100 00:05:56.531 tick 100 00:05:56.531 tick 250 00:05:56.531 tick 500 00:05:56.531 tick 100 00:05:56.531 tick 100 00:05:56.532 tick 250 00:05:56.532 tick 100 00:05:56.532 tick 100 00:05:56.532 test_end 00:05:56.532 00:05:56.532 real 0m1.347s 00:05:56.532 user 0m1.137s 00:05:56.532 sys 0m0.104s 00:05:56.532 14:05:01 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:56.532 14:05:01 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:05:56.532 ************************************ 00:05:56.532 END TEST event_reactor 00:05:56.532 ************************************ 00:05:56.532 14:05:01 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:56.532 14:05:01 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:05:56.532 14:05:01 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:56.532 14:05:01 event -- common/autotest_common.sh@10 -- # set +x 00:05:56.792 ************************************ 00:05:56.792 START TEST event_reactor_perf 00:05:56.792 ************************************ 00:05:56.792 14:05:01 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:56.792 [2024-09-30 14:05:01.237644] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:56.792 [2024-09-30 14:05:01.237765] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71220 ] 00:05:56.792 [2024-09-30 14:05:01.371248] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:56.792 [2024-09-30 14:05:01.400292] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:56.792 [2024-09-30 14:05:01.443673] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:58.172 test_start 00:05:58.172 test_end 00:05:58.172 Performance: 410720 events per second 00:05:58.172 00:05:58.172 real 0m1.345s 00:05:58.172 user 0m1.139s 00:05:58.172 sys 0m0.100s 00:05:58.172 14:05:02 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:58.172 ************************************ 00:05:58.172 END TEST event_reactor_perf 00:05:58.172 ************************************ 00:05:58.172 14:05:02 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:05:58.172 14:05:02 event -- event/event.sh@49 -- # uname -s 00:05:58.172 14:05:02 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:05:58.172 14:05:02 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:58.172 14:05:02 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:58.172 14:05:02 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:58.172 14:05:02 event -- common/autotest_common.sh@10 -- # set +x 00:05:58.172 ************************************ 00:05:58.172 START TEST event_scheduler 00:05:58.172 ************************************ 00:05:58.172 14:05:02 event.event_scheduler -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:58.172 * Looking for test storage... 00:05:58.172 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:05:58.172 14:05:02 event.event_scheduler -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:58.172 14:05:02 event.event_scheduler -- common/autotest_common.sh@1681 -- # lcov --version 00:05:58.172 14:05:02 event.event_scheduler -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:58.172 14:05:02 event.event_scheduler -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:58.172 14:05:02 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:58.432 14:05:02 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:58.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:58.432 --rc genhtml_branch_coverage=1 00:05:58.432 --rc genhtml_function_coverage=1 00:05:58.432 --rc genhtml_legend=1 00:05:58.432 --rc geninfo_all_blocks=1 00:05:58.432 --rc geninfo_unexecuted_blocks=1 00:05:58.432 00:05:58.432 ' 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:58.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:58.432 --rc genhtml_branch_coverage=1 00:05:58.432 --rc genhtml_function_coverage=1 00:05:58.432 --rc genhtml_legend=1 00:05:58.432 --rc geninfo_all_blocks=1 00:05:58.432 --rc geninfo_unexecuted_blocks=1 00:05:58.432 00:05:58.432 ' 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:58.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:58.432 --rc genhtml_branch_coverage=1 00:05:58.432 --rc genhtml_function_coverage=1 00:05:58.432 --rc genhtml_legend=1 00:05:58.432 --rc geninfo_all_blocks=1 00:05:58.432 --rc geninfo_unexecuted_blocks=1 00:05:58.432 00:05:58.432 ' 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:58.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:58.432 --rc genhtml_branch_coverage=1 00:05:58.432 --rc genhtml_function_coverage=1 00:05:58.432 --rc genhtml_legend=1 00:05:58.432 --rc geninfo_all_blocks=1 00:05:58.432 --rc geninfo_unexecuted_blocks=1 00:05:58.432 00:05:58.432 ' 00:05:58.432 14:05:02 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:05:58.432 14:05:02 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=71296 00:05:58.432 14:05:02 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:05:58.432 14:05:02 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:05:58.432 14:05:02 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 71296 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 71296 ']' 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:58.432 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:58.432 14:05:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:58.432 [2024-09-30 14:05:02.906063] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:05:58.432 [2024-09-30 14:05:02.906598] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71296 ] 00:05:58.432 [2024-09-30 14:05:03.041809] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:05:58.432 [2024-09-30 14:05:03.069854] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:58.692 [2024-09-30 14:05:03.117981] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:58.692 [2024-09-30 14:05:03.118177] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:58.692 [2024-09-30 14:05:03.118672] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:05:58.692 [2024-09-30 14:05:03.118749] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:05:59.262 14:05:03 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:59.262 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:59.262 POWER: Cannot set governor of lcore 0 to userspace 00:05:59.262 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:59.262 POWER: Cannot set governor of lcore 0 to performance 00:05:59.262 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:59.262 POWER: Cannot set governor of lcore 0 to userspace 00:05:59.262 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:59.262 POWER: Cannot set governor of lcore 0 to userspace 00:05:59.262 GUEST_CHANNEL: Opening channel '/dev/virtio-ports/virtio.serial.port.poweragent.0' for lcore 0 00:05:59.262 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:05:59.262 POWER: Unable to set Power Management Environment for lcore 0 00:05:59.262 [2024-09-30 14:05:03.759888] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:05:59.262 [2024-09-30 14:05:03.759937] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:05:59.262 [2024-09-30 14:05:03.759968] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:05:59.262 [2024-09-30 14:05:03.760005] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:05:59.262 [2024-09-30 14:05:03.760062] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:05:59.262 [2024-09-30 14:05:03.760116] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.262 14:05:03 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:59.262 [2024-09-30 14:05:03.830900] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.262 14:05:03 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:59.262 14:05:03 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:59.262 ************************************ 00:05:59.262 START TEST scheduler_create_thread 00:05:59.262 ************************************ 00:05:59.262 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:05:59.262 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 2 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 3 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 4 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 5 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 6 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 7 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.263 8 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.263 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.524 9 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.524 10 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.524 14:05:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:00.463 14:05:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:00.463 14:05:04 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:00.463 14:05:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:00.463 14:05:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:01.844 14:05:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:01.844 14:05:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:01.844 14:05:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:01.844 14:05:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:01.844 14:05:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:02.783 ************************************ 00:06:02.783 END TEST scheduler_create_thread 00:06:02.783 ************************************ 00:06:02.783 14:05:07 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:02.783 00:06:02.783 real 0m3.375s 00:06:02.783 user 0m0.030s 00:06:02.783 sys 0m0.006s 00:06:02.783 14:05:07 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:02.783 14:05:07 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:02.783 14:05:07 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:02.783 14:05:07 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 71296 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 71296 ']' 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 71296 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71296 00:06:02.783 killing process with pid 71296 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71296' 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 71296 00:06:02.783 14:05:07 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 71296 00:06:03.043 [2024-09-30 14:05:07.595861] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:03.302 00:06:03.303 real 0m5.270s 00:06:03.303 user 0m10.329s 00:06:03.303 sys 0m0.497s 00:06:03.303 ************************************ 00:06:03.303 END TEST event_scheduler 00:06:03.303 14:05:07 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:03.303 14:05:07 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:03.303 ************************************ 00:06:03.303 14:05:07 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:03.303 14:05:07 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:03.303 14:05:07 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:03.303 14:05:07 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:03.303 14:05:07 event -- common/autotest_common.sh@10 -- # set +x 00:06:03.563 ************************************ 00:06:03.563 START TEST app_repeat 00:06:03.563 ************************************ 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@19 -- # repeat_pid=71397 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 71397' 00:06:03.563 Process app_repeat pid: 71397 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:03.563 spdk_app_start Round 0 00:06:03.563 14:05:07 event.app_repeat -- event/event.sh@25 -- # waitforlisten 71397 /var/tmp/spdk-nbd.sock 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 71397 ']' 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:03.563 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:03.563 14:05:07 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:03.563 [2024-09-30 14:05:08.023865] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:03.563 [2024-09-30 14:05:08.024026] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71397 ] 00:06:03.563 [2024-09-30 14:05:08.155770] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:03.563 [2024-09-30 14:05:08.185965] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:03.823 [2024-09-30 14:05:08.232361] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:03.823 [2024-09-30 14:05:08.232464] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:04.394 14:05:08 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:04.394 14:05:08 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:04.394 14:05:08 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:04.394 Malloc0 00:06:04.653 14:05:09 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:04.653 Malloc1 00:06:04.653 14:05:09 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:04.653 14:05:09 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:04.912 /dev/nbd0 00:06:04.912 14:05:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:04.912 14:05:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:04.912 1+0 records in 00:06:04.912 1+0 records out 00:06:04.912 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00036934 s, 11.1 MB/s 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:04.912 14:05:09 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:04.912 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:04.912 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:04.912 14:05:09 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:05.172 /dev/nbd1 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:05.172 1+0 records in 00:06:05.172 1+0 records out 00:06:05.172 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000397374 s, 10.3 MB/s 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:05.172 14:05:09 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:05.172 14:05:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:05.433 { 00:06:05.433 "nbd_device": "/dev/nbd0", 00:06:05.433 "bdev_name": "Malloc0" 00:06:05.433 }, 00:06:05.433 { 00:06:05.433 "nbd_device": "/dev/nbd1", 00:06:05.433 "bdev_name": "Malloc1" 00:06:05.433 } 00:06:05.433 ]' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:05.433 { 00:06:05.433 "nbd_device": "/dev/nbd0", 00:06:05.433 "bdev_name": "Malloc0" 00:06:05.433 }, 00:06:05.433 { 00:06:05.433 "nbd_device": "/dev/nbd1", 00:06:05.433 "bdev_name": "Malloc1" 00:06:05.433 } 00:06:05.433 ]' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:05.433 /dev/nbd1' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:05.433 /dev/nbd1' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:05.433 256+0 records in 00:06:05.433 256+0 records out 00:06:05.433 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0123577 s, 84.9 MB/s 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:05.433 14:05:09 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:05.433 256+0 records in 00:06:05.433 256+0 records out 00:06:05.433 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0241108 s, 43.5 MB/s 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:05.433 256+0 records in 00:06:05.433 256+0 records out 00:06:05.433 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0293014 s, 35.8 MB/s 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:05.433 14:05:10 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:05.693 14:05:10 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:05.954 14:05:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:06.214 14:05:10 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:06.214 14:05:10 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:06.474 14:05:10 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:06.474 [2024-09-30 14:05:11.108600] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:06.733 [2024-09-30 14:05:11.151400] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:06.733 [2024-09-30 14:05:11.151403] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:06.733 [2024-09-30 14:05:11.192780] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:06.733 [2024-09-30 14:05:11.192898] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:10.024 spdk_app_start Round 1 00:06:10.024 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:10.024 14:05:13 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:10.024 14:05:13 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:10.024 14:05:13 event.app_repeat -- event/event.sh@25 -- # waitforlisten 71397 /var/tmp/spdk-nbd.sock 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 71397 ']' 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:10.024 14:05:13 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:10.024 14:05:14 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:10.024 14:05:14 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:10.024 14:05:14 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:10.024 Malloc0 00:06:10.024 14:05:14 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:10.024 Malloc1 00:06:10.024 14:05:14 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:10.024 14:05:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:10.283 /dev/nbd0 00:06:10.283 14:05:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:10.283 14:05:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:10.283 1+0 records in 00:06:10.283 1+0 records out 00:06:10.283 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000263158 s, 15.6 MB/s 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:10.283 14:05:14 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:10.283 14:05:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:10.283 14:05:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:10.283 14:05:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:10.542 /dev/nbd1 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:10.542 1+0 records in 00:06:10.542 1+0 records out 00:06:10.542 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000357462 s, 11.5 MB/s 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:10.542 14:05:15 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:10.542 14:05:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:10.802 { 00:06:10.802 "nbd_device": "/dev/nbd0", 00:06:10.802 "bdev_name": "Malloc0" 00:06:10.802 }, 00:06:10.802 { 00:06:10.802 "nbd_device": "/dev/nbd1", 00:06:10.802 "bdev_name": "Malloc1" 00:06:10.802 } 00:06:10.802 ]' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:10.802 { 00:06:10.802 "nbd_device": "/dev/nbd0", 00:06:10.802 "bdev_name": "Malloc0" 00:06:10.802 }, 00:06:10.802 { 00:06:10.802 "nbd_device": "/dev/nbd1", 00:06:10.802 "bdev_name": "Malloc1" 00:06:10.802 } 00:06:10.802 ]' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:10.802 /dev/nbd1' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:10.802 /dev/nbd1' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:10.802 256+0 records in 00:06:10.802 256+0 records out 00:06:10.802 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0130387 s, 80.4 MB/s 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:10.802 256+0 records in 00:06:10.802 256+0 records out 00:06:10.802 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0260906 s, 40.2 MB/s 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:10.802 256+0 records in 00:06:10.802 256+0 records out 00:06:10.802 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0233924 s, 44.8 MB/s 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:10.802 14:05:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:11.062 14:05:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:11.324 14:05:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:11.592 14:05:16 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:11.592 14:05:16 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:11.857 14:05:16 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:11.857 [2024-09-30 14:05:16.480503] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:12.115 [2024-09-30 14:05:16.522421] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:12.115 [2024-09-30 14:05:16.522475] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:12.115 [2024-09-30 14:05:16.562942] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:12.115 [2024-09-30 14:05:16.563013] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:15.406 spdk_app_start Round 2 00:06:15.406 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@25 -- # waitforlisten 71397 /var/tmp/spdk-nbd.sock 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 71397 ']' 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:15.406 14:05:19 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:15.406 Malloc0 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:15.406 Malloc1 00:06:15.406 14:05:19 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:15.406 14:05:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:15.666 /dev/nbd0 00:06:15.666 14:05:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:15.666 14:05:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:15.666 1+0 records in 00:06:15.666 1+0 records out 00:06:15.666 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000368846 s, 11.1 MB/s 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:15.666 14:05:20 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:15.666 14:05:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:15.666 14:05:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:15.666 14:05:20 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:15.925 /dev/nbd1 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:15.925 1+0 records in 00:06:15.925 1+0 records out 00:06:15.925 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000376137 s, 10.9 MB/s 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:15.925 14:05:20 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:15.925 14:05:20 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:15.926 14:05:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:16.185 14:05:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:16.185 { 00:06:16.185 "nbd_device": "/dev/nbd0", 00:06:16.185 "bdev_name": "Malloc0" 00:06:16.185 }, 00:06:16.185 { 00:06:16.185 "nbd_device": "/dev/nbd1", 00:06:16.186 "bdev_name": "Malloc1" 00:06:16.186 } 00:06:16.186 ]' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:16.186 { 00:06:16.186 "nbd_device": "/dev/nbd0", 00:06:16.186 "bdev_name": "Malloc0" 00:06:16.186 }, 00:06:16.186 { 00:06:16.186 "nbd_device": "/dev/nbd1", 00:06:16.186 "bdev_name": "Malloc1" 00:06:16.186 } 00:06:16.186 ]' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:16.186 /dev/nbd1' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:16.186 /dev/nbd1' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:16.186 256+0 records in 00:06:16.186 256+0 records out 00:06:16.186 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00642472 s, 163 MB/s 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:16.186 256+0 records in 00:06:16.186 256+0 records out 00:06:16.186 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0251197 s, 41.7 MB/s 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:16.186 256+0 records in 00:06:16.186 256+0 records out 00:06:16.186 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0245734 s, 42.7 MB/s 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:16.186 14:05:20 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:16.445 14:05:21 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:16.705 14:05:21 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:16.706 14:05:21 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:16.706 14:05:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:16.966 14:05:21 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:16.966 14:05:21 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:17.226 14:05:21 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:17.226 [2024-09-30 14:05:21.830463] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:17.226 [2024-09-30 14:05:21.872086] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:17.226 [2024-09-30 14:05:21.872093] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:17.486 [2024-09-30 14:05:21.913130] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:17.486 [2024-09-30 14:05:21.913191] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:20.023 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:20.023 14:05:24 event.app_repeat -- event/event.sh@38 -- # waitforlisten 71397 /var/tmp/spdk-nbd.sock 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 71397 ']' 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:20.023 14:05:24 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:20.283 14:05:24 event.app_repeat -- event/event.sh@39 -- # killprocess 71397 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 71397 ']' 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 71397 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71397 00:06:20.283 killing process with pid 71397 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71397' 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@969 -- # kill 71397 00:06:20.283 14:05:24 event.app_repeat -- common/autotest_common.sh@974 -- # wait 71397 00:06:20.543 spdk_app_start is called in Round 0. 00:06:20.543 Shutdown signal received, stop current app iteration 00:06:20.543 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 reinitialization... 00:06:20.543 spdk_app_start is called in Round 1. 00:06:20.543 Shutdown signal received, stop current app iteration 00:06:20.543 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 reinitialization... 00:06:20.543 spdk_app_start is called in Round 2. 00:06:20.543 Shutdown signal received, stop current app iteration 00:06:20.543 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 reinitialization... 00:06:20.543 spdk_app_start is called in Round 3. 00:06:20.543 Shutdown signal received, stop current app iteration 00:06:20.543 ************************************ 00:06:20.543 END TEST app_repeat 00:06:20.543 ************************************ 00:06:20.543 14:05:25 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:20.543 14:05:25 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:20.543 00:06:20.543 real 0m17.149s 00:06:20.543 user 0m37.863s 00:06:20.543 sys 0m2.280s 00:06:20.543 14:05:25 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:20.543 14:05:25 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:20.543 14:05:25 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:20.543 14:05:25 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:20.543 14:05:25 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:20.543 14:05:25 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:20.543 14:05:25 event -- common/autotest_common.sh@10 -- # set +x 00:06:20.543 ************************************ 00:06:20.543 START TEST cpu_locks 00:06:20.544 ************************************ 00:06:20.544 14:05:25 event.cpu_locks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:20.804 * Looking for test storage... 00:06:20.804 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1681 -- # lcov --version 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:20.804 14:05:25 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:20.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.804 --rc genhtml_branch_coverage=1 00:06:20.804 --rc genhtml_function_coverage=1 00:06:20.804 --rc genhtml_legend=1 00:06:20.804 --rc geninfo_all_blocks=1 00:06:20.804 --rc geninfo_unexecuted_blocks=1 00:06:20.804 00:06:20.804 ' 00:06:20.804 14:05:25 event.cpu_locks -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:20.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.805 --rc genhtml_branch_coverage=1 00:06:20.805 --rc genhtml_function_coverage=1 00:06:20.805 --rc genhtml_legend=1 00:06:20.805 --rc geninfo_all_blocks=1 00:06:20.805 --rc geninfo_unexecuted_blocks=1 00:06:20.805 00:06:20.805 ' 00:06:20.805 14:05:25 event.cpu_locks -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:20.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.805 --rc genhtml_branch_coverage=1 00:06:20.805 --rc genhtml_function_coverage=1 00:06:20.805 --rc genhtml_legend=1 00:06:20.805 --rc geninfo_all_blocks=1 00:06:20.805 --rc geninfo_unexecuted_blocks=1 00:06:20.805 00:06:20.805 ' 00:06:20.805 14:05:25 event.cpu_locks -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:20.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.805 --rc genhtml_branch_coverage=1 00:06:20.805 --rc genhtml_function_coverage=1 00:06:20.805 --rc genhtml_legend=1 00:06:20.805 --rc geninfo_all_blocks=1 00:06:20.805 --rc geninfo_unexecuted_blocks=1 00:06:20.805 00:06:20.805 ' 00:06:20.805 14:05:25 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:20.805 14:05:25 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:20.805 14:05:25 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:20.805 14:05:25 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:20.805 14:05:25 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:20.805 14:05:25 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:20.805 14:05:25 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:20.805 ************************************ 00:06:20.805 START TEST default_locks 00:06:20.805 ************************************ 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=71823 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 71823 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 71823 ']' 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:20.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:20.805 14:05:25 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:21.065 [2024-09-30 14:05:25.492318] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:21.065 [2024-09-30 14:05:25.492448] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71823 ] 00:06:21.065 [2024-09-30 14:05:25.622597] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:21.065 [2024-09-30 14:05:25.651096] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:21.065 [2024-09-30 14:05:25.696067] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.003 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:22.003 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:06:22.003 14:05:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 71823 00:06:22.003 14:05:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:22.003 14:05:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 71823 00:06:22.262 14:05:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 71823 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 71823 ']' 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 71823 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71823 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:22.263 killing process with pid 71823 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71823' 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 71823 00:06:22.263 14:05:26 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 71823 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 71823 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 71823 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 71823 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 71823 ']' 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:22.523 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:22.523 ERROR: process (pid: 71823) is no longer running 00:06:22.523 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (71823) - No such process 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:22.523 14:05:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:22.524 14:05:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:22.524 14:05:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:22.524 00:06:22.524 real 0m1.725s 00:06:22.524 user 0m1.667s 00:06:22.524 sys 0m0.625s 00:06:22.524 ************************************ 00:06:22.524 END TEST default_locks 00:06:22.524 ************************************ 00:06:22.524 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:22.524 14:05:27 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:22.524 14:05:27 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:22.524 14:05:27 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:22.524 14:05:27 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:22.524 14:05:27 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:22.783 ************************************ 00:06:22.783 START TEST default_locks_via_rpc 00:06:22.783 ************************************ 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=71871 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 71871 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 71871 ']' 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:22.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:22.783 14:05:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:22.783 [2024-09-30 14:05:27.280191] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:22.783 [2024-09-30 14:05:27.280387] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71871 ] 00:06:22.783 [2024-09-30 14:05:27.410175] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:22.783 [2024-09-30 14:05:27.429038] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:23.043 [2024-09-30 14:05:27.477696] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 71871 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 71871 00:06:23.611 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 71871 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 71871 ']' 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 71871 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71871 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71871' 00:06:24.181 killing process with pid 71871 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 71871 00:06:24.181 14:05:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 71871 00:06:24.441 00:06:24.441 real 0m1.814s 00:06:24.441 user 0m1.815s 00:06:24.441 sys 0m0.604s 00:06:24.441 14:05:29 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:24.441 14:05:29 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:24.441 ************************************ 00:06:24.441 END TEST default_locks_via_rpc 00:06:24.441 ************************************ 00:06:24.441 14:05:29 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:24.441 14:05:29 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:24.441 14:05:29 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:24.441 14:05:29 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:24.442 ************************************ 00:06:24.442 START TEST non_locking_app_on_locked_coremask 00:06:24.442 ************************************ 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=71922 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 71922 /var/tmp/spdk.sock 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 71922 ']' 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:24.442 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:24.442 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:24.702 [2024-09-30 14:05:29.173940] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:24.702 [2024-09-30 14:05:29.174077] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71922 ] 00:06:24.702 [2024-09-30 14:05:29.308898] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:24.702 [2024-09-30 14:05:29.338163] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:24.961 [2024-09-30 14:05:29.383710] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=71934 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 71934 /var/tmp/spdk2.sock 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 71934 ']' 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:25.531 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:25.531 14:05:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:25.531 [2024-09-30 14:05:30.042184] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:25.531 [2024-09-30 14:05:30.042439] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71934 ] 00:06:25.531 [2024-09-30 14:05:30.181170] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:25.790 [2024-09-30 14:05:30.199047] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:25.791 [2024-09-30 14:05:30.199087] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:25.791 [2024-09-30 14:05:30.289005] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:26.359 14:05:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:26.359 14:05:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:26.359 14:05:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 71922 00:06:26.359 14:05:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71922 00:06:26.359 14:05:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 71922 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 71922 ']' 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 71922 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:26.618 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71922 00:06:26.877 killing process with pid 71922 00:06:26.877 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:26.877 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:26.877 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71922' 00:06:26.877 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 71922 00:06:26.877 14:05:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 71922 00:06:27.447 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 71934 00:06:27.447 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 71934 ']' 00:06:27.447 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 71934 00:06:27.447 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:27.448 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:27.448 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71934 00:06:27.747 killing process with pid 71934 00:06:27.747 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:27.747 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:27.747 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71934' 00:06:27.747 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 71934 00:06:27.747 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 71934 00:06:28.006 00:06:28.006 real 0m3.428s 00:06:28.006 user 0m3.583s 00:06:28.006 sys 0m1.027s 00:06:28.006 ************************************ 00:06:28.006 END TEST non_locking_app_on_locked_coremask 00:06:28.006 ************************************ 00:06:28.006 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:28.006 14:05:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:28.006 14:05:32 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:28.006 14:05:32 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:28.006 14:05:32 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:28.006 14:05:32 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:28.006 ************************************ 00:06:28.006 START TEST locking_app_on_unlocked_coremask 00:06:28.006 ************************************ 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=71998 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 71998 /var/tmp/spdk.sock 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 71998 ']' 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:28.006 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:28.006 14:05:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:28.006 [2024-09-30 14:05:32.658229] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:28.006 [2024-09-30 14:05:32.658418] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71998 ] 00:06:28.265 [2024-09-30 14:05:32.790286] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:28.265 [2024-09-30 14:05:32.818485] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:28.265 [2024-09-30 14:05:32.818549] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:28.265 [2024-09-30 14:05:32.866254] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=72014 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 72014 /var/tmp/spdk2.sock 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 72014 ']' 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:28.832 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:28.832 14:05:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:29.092 [2024-09-30 14:05:33.563849] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:29.092 [2024-09-30 14:05:33.564061] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72014 ] 00:06:29.092 [2024-09-30 14:05:33.693164] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:29.092 [2024-09-30 14:05:33.714596] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:29.351 [2024-09-30 14:05:33.800070] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:29.919 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:29.919 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:29.919 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 72014 00:06:29.919 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:29.919 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 72014 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 71998 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 71998 ']' 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 71998 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71998 00:06:30.179 killing process with pid 71998 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71998' 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 71998 00:06:30.179 14:05:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 71998 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 72014 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 72014 ']' 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 72014 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72014 00:06:31.117 killing process with pid 72014 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72014' 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 72014 00:06:31.117 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 72014 00:06:31.377 00:06:31.377 real 0m3.379s 00:06:31.377 user 0m3.581s 00:06:31.377 sys 0m0.980s 00:06:31.377 ************************************ 00:06:31.377 END TEST locking_app_on_unlocked_coremask 00:06:31.377 ************************************ 00:06:31.377 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:31.377 14:05:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:31.377 14:05:35 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:31.377 14:05:35 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:31.377 14:05:35 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:31.377 14:05:35 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:31.377 ************************************ 00:06:31.377 START TEST locking_app_on_locked_coremask 00:06:31.377 ************************************ 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=72078 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 72078 /var/tmp/spdk.sock 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 72078 ']' 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:31.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:31.377 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:31.636 [2024-09-30 14:05:36.095345] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:31.636 [2024-09-30 14:05:36.095468] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72078 ] 00:06:31.636 [2024-09-30 14:05:36.225717] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:31.636 [2024-09-30 14:05:36.254918] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:31.896 [2024-09-30 14:05:36.300740] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=72094 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 72094 /var/tmp/spdk2.sock 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 72094 /var/tmp/spdk2.sock 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:32.465 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 72094 /var/tmp/spdk2.sock 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 72094 ']' 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:32.466 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:32.466 14:05:36 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:32.466 [2024-09-30 14:05:37.005552] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:32.466 [2024-09-30 14:05:37.005773] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72094 ] 00:06:32.725 [2024-09-30 14:05:37.138112] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:32.725 [2024-09-30 14:05:37.155770] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 72078 has claimed it. 00:06:32.725 [2024-09-30 14:05:37.155840] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:32.984 ERROR: process (pid: 72094) is no longer running 00:06:32.984 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (72094) - No such process 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 72078 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:32.984 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 72078 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 72078 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 72078 ']' 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 72078 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:33.551 14:05:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72078 00:06:33.551 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:33.551 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:33.551 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72078' 00:06:33.551 killing process with pid 72078 00:06:33.551 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 72078 00:06:33.551 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 72078 00:06:33.811 00:06:33.811 real 0m2.407s 00:06:33.811 user 0m2.584s 00:06:33.811 sys 0m0.721s 00:06:33.811 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:33.811 14:05:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:33.811 ************************************ 00:06:33.811 END TEST locking_app_on_locked_coremask 00:06:33.811 ************************************ 00:06:33.811 14:05:38 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:33.811 14:05:38 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:33.811 14:05:38 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:33.811 14:05:38 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:34.071 ************************************ 00:06:34.071 START TEST locking_overlapped_coremask 00:06:34.071 ************************************ 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=72136 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 72136 /var/tmp/spdk.sock 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 72136 ']' 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:34.071 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:34.071 14:05:38 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:34.071 [2024-09-30 14:05:38.565496] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:34.071 [2024-09-30 14:05:38.565680] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72136 ] 00:06:34.071 [2024-09-30 14:05:38.697647] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:34.071 [2024-09-30 14:05:38.724461] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:34.331 [2024-09-30 14:05:38.774123] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:34.331 [2024-09-30 14:05:38.774218] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:34.331 [2024-09-30 14:05:38.774335] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=72154 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 72154 /var/tmp/spdk2.sock 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 72154 /var/tmp/spdk2.sock 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 72154 /var/tmp/spdk2.sock 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 72154 ']' 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:34.902 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:34.902 14:05:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:34.902 [2024-09-30 14:05:39.473892] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:34.902 [2024-09-30 14:05:39.474085] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72154 ] 00:06:35.162 [2024-09-30 14:05:39.610671] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:35.162 [2024-09-30 14:05:39.633189] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 72136 has claimed it. 00:06:35.162 [2024-09-30 14:05:39.633254] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:35.733 ERROR: process (pid: 72154) is no longer running 00:06:35.733 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (72154) - No such process 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 72136 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 72136 ']' 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 72136 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72136 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72136' 00:06:35.733 killing process with pid 72136 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 72136 00:06:35.733 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 72136 00:06:35.992 00:06:35.992 real 0m2.073s 00:06:35.992 user 0m5.470s 00:06:35.992 sys 0m0.542s 00:06:35.992 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:35.992 14:05:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:35.993 ************************************ 00:06:35.993 END TEST locking_overlapped_coremask 00:06:35.993 ************************************ 00:06:35.993 14:05:40 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:35.993 14:05:40 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:35.993 14:05:40 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:35.993 14:05:40 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:35.993 ************************************ 00:06:35.993 START TEST locking_overlapped_coremask_via_rpc 00:06:35.993 ************************************ 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=72196 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 72196 /var/tmp/spdk.sock 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 72196 ']' 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:35.993 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:35.993 14:05:40 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:36.252 [2024-09-30 14:05:40.709827] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:36.252 [2024-09-30 14:05:40.710017] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72196 ] 00:06:36.252 [2024-09-30 14:05:40.841806] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:36.252 [2024-09-30 14:05:40.860445] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:36.252 [2024-09-30 14:05:40.860548] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:36.252 [2024-09-30 14:05:40.907206] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:36.252 [2024-09-30 14:05:40.907433] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:36.252 [2024-09-30 14:05:40.907319] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=72214 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 72214 /var/tmp/spdk2.sock 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 72214 ']' 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:37.189 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:37.189 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:37.190 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:37.190 14:05:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.190 [2024-09-30 14:05:41.585216] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:37.190 [2024-09-30 14:05:41.585345] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72214 ] 00:06:37.190 [2024-09-30 14:05:41.720609] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:37.190 [2024-09-30 14:05:41.738957] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:37.190 [2024-09-30 14:05:41.738996] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:37.448 [2024-09-30 14:05:41.905334] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:06:37.449 [2024-09-30 14:05:41.905445] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:37.449 [2024-09-30 14:05:41.905580] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 4 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.017 [2024-09-30 14:05:42.616719] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 72196 has claimed it. 00:06:38.017 request: 00:06:38.017 { 00:06:38.017 "method": "framework_enable_cpumask_locks", 00:06:38.017 "req_id": 1 00:06:38.017 } 00:06:38.017 Got JSON-RPC error response 00:06:38.017 response: 00:06:38.017 { 00:06:38.017 "code": -32603, 00:06:38.017 "message": "Failed to claim CPU core: 2" 00:06:38.017 } 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 72196 /var/tmp/spdk.sock 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 72196 ']' 00:06:38.017 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:38.018 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:38.018 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:38.018 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:38.018 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:38.018 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 72214 /var/tmp/spdk2.sock 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 72214 ']' 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:38.277 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:38.277 14:05:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:38.537 00:06:38.537 real 0m2.429s 00:06:38.537 user 0m1.028s 00:06:38.537 sys 0m0.178s 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:38.537 14:05:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.537 ************************************ 00:06:38.537 END TEST locking_overlapped_coremask_via_rpc 00:06:38.537 ************************************ 00:06:38.537 14:05:43 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:38.537 14:05:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 72196 ]] 00:06:38.537 14:05:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 72196 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 72196 ']' 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 72196 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72196 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72196' 00:06:38.537 killing process with pid 72196 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 72196 00:06:38.537 14:05:43 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 72196 00:06:39.107 14:05:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 72214 ]] 00:06:39.107 14:05:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 72214 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 72214 ']' 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 72214 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72214 00:06:39.107 killing process with pid 72214 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72214' 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 72214 00:06:39.107 14:05:43 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 72214 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:39.677 Process with pid 72196 is not found 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 72196 ]] 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 72196 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 72196 ']' 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 72196 00:06:39.677 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (72196) - No such process 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 72196 is not found' 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 72214 ]] 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 72214 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 72214 ']' 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 72214 00:06:39.677 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (72214) - No such process 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 72214 is not found' 00:06:39.677 Process with pid 72214 is not found 00:06:39.677 14:05:44 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:39.677 ************************************ 00:06:39.677 END TEST cpu_locks 00:06:39.677 ************************************ 00:06:39.677 00:06:39.677 real 0m19.114s 00:06:39.677 user 0m32.304s 00:06:39.677 sys 0m5.977s 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:39.677 14:05:44 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:39.936 ************************************ 00:06:39.936 END TEST event 00:06:39.936 ************************************ 00:06:39.936 00:06:39.936 real 0m46.229s 00:06:39.936 user 1m27.157s 00:06:39.936 sys 0m9.466s 00:06:39.936 14:05:44 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:39.936 14:05:44 event -- common/autotest_common.sh@10 -- # set +x 00:06:39.936 14:05:44 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:39.936 14:05:44 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:39.936 14:05:44 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:39.936 14:05:44 -- common/autotest_common.sh@10 -- # set +x 00:06:39.936 ************************************ 00:06:39.936 START TEST thread 00:06:39.936 ************************************ 00:06:39.936 14:05:44 thread -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:39.936 * Looking for test storage... 00:06:39.936 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:39.936 14:05:44 thread -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:39.936 14:05:44 thread -- common/autotest_common.sh@1681 -- # lcov --version 00:06:39.936 14:05:44 thread -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:40.196 14:05:44 thread -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:40.196 14:05:44 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:40.196 14:05:44 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:40.196 14:05:44 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:40.196 14:05:44 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:40.196 14:05:44 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:40.196 14:05:44 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:40.196 14:05:44 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:40.196 14:05:44 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:40.196 14:05:44 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:40.196 14:05:44 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:40.196 14:05:44 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:40.197 14:05:44 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:40.197 14:05:44 thread -- scripts/common.sh@345 -- # : 1 00:06:40.197 14:05:44 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:40.197 14:05:44 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:40.197 14:05:44 thread -- scripts/common.sh@365 -- # decimal 1 00:06:40.197 14:05:44 thread -- scripts/common.sh@353 -- # local d=1 00:06:40.197 14:05:44 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:40.197 14:05:44 thread -- scripts/common.sh@355 -- # echo 1 00:06:40.197 14:05:44 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:40.197 14:05:44 thread -- scripts/common.sh@366 -- # decimal 2 00:06:40.197 14:05:44 thread -- scripts/common.sh@353 -- # local d=2 00:06:40.197 14:05:44 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:40.197 14:05:44 thread -- scripts/common.sh@355 -- # echo 2 00:06:40.197 14:05:44 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:40.197 14:05:44 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:40.197 14:05:44 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:40.197 14:05:44 thread -- scripts/common.sh@368 -- # return 0 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:40.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.197 --rc genhtml_branch_coverage=1 00:06:40.197 --rc genhtml_function_coverage=1 00:06:40.197 --rc genhtml_legend=1 00:06:40.197 --rc geninfo_all_blocks=1 00:06:40.197 --rc geninfo_unexecuted_blocks=1 00:06:40.197 00:06:40.197 ' 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:40.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.197 --rc genhtml_branch_coverage=1 00:06:40.197 --rc genhtml_function_coverage=1 00:06:40.197 --rc genhtml_legend=1 00:06:40.197 --rc geninfo_all_blocks=1 00:06:40.197 --rc geninfo_unexecuted_blocks=1 00:06:40.197 00:06:40.197 ' 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:40.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.197 --rc genhtml_branch_coverage=1 00:06:40.197 --rc genhtml_function_coverage=1 00:06:40.197 --rc genhtml_legend=1 00:06:40.197 --rc geninfo_all_blocks=1 00:06:40.197 --rc geninfo_unexecuted_blocks=1 00:06:40.197 00:06:40.197 ' 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:40.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.197 --rc genhtml_branch_coverage=1 00:06:40.197 --rc genhtml_function_coverage=1 00:06:40.197 --rc genhtml_legend=1 00:06:40.197 --rc geninfo_all_blocks=1 00:06:40.197 --rc geninfo_unexecuted_blocks=1 00:06:40.197 00:06:40.197 ' 00:06:40.197 14:05:44 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:40.197 14:05:44 thread -- common/autotest_common.sh@10 -- # set +x 00:06:40.197 ************************************ 00:06:40.197 START TEST thread_poller_perf 00:06:40.197 ************************************ 00:06:40.197 14:05:44 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:40.197 [2024-09-30 14:05:44.692252] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:40.197 [2024-09-30 14:05:44.692382] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72352 ] 00:06:40.197 [2024-09-30 14:05:44.822572] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:40.457 [2024-09-30 14:05:44.854640] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:40.457 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:40.457 [2024-09-30 14:05:44.902446] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:41.395 ====================================== 00:06:41.395 busy:2299648378 (cyc) 00:06:41.395 total_run_count: 427000 00:06:41.395 tsc_hz: 2290000000 (cyc) 00:06:41.395 ====================================== 00:06:41.395 poller_cost: 5385 (cyc), 2351 (nsec) 00:06:41.395 00:06:41.395 real 0m1.347s 00:06:41.395 user 0m1.161s 00:06:41.395 sys 0m0.080s 00:06:41.395 ************************************ 00:06:41.395 END TEST thread_poller_perf 00:06:41.395 ************************************ 00:06:41.395 14:05:45 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:41.395 14:05:45 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:41.395 14:05:46 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:41.395 14:05:46 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:06:41.395 14:05:46 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:41.396 14:05:46 thread -- common/autotest_common.sh@10 -- # set +x 00:06:41.655 ************************************ 00:06:41.655 START TEST thread_poller_perf 00:06:41.655 ************************************ 00:06:41.655 14:05:46 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:41.655 [2024-09-30 14:05:46.100585] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:41.655 [2024-09-30 14:05:46.100764] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72396 ] 00:06:41.655 [2024-09-30 14:05:46.230664] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:41.655 [2024-09-30 14:05:46.259425] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:41.656 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:41.656 [2024-09-30 14:05:46.304144] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:43.036 ====================================== 00:06:43.036 busy:2293055388 (cyc) 00:06:43.036 total_run_count: 5569000 00:06:43.036 tsc_hz: 2290000000 (cyc) 00:06:43.036 ====================================== 00:06:43.036 poller_cost: 411 (cyc), 179 (nsec) 00:06:43.036 00:06:43.036 real 0m1.343s 00:06:43.036 user 0m1.138s 00:06:43.036 sys 0m0.099s 00:06:43.036 14:05:47 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:43.036 ************************************ 00:06:43.036 END TEST thread_poller_perf 00:06:43.036 ************************************ 00:06:43.036 14:05:47 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:43.036 14:05:47 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:06:43.036 ************************************ 00:06:43.036 END TEST thread 00:06:43.036 ************************************ 00:06:43.036 00:06:43.036 real 0m3.046s 00:06:43.036 user 0m2.467s 00:06:43.036 sys 0m0.382s 00:06:43.036 14:05:47 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:43.036 14:05:47 thread -- common/autotest_common.sh@10 -- # set +x 00:06:43.036 14:05:47 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:06:43.036 14:05:47 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:43.036 14:05:47 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:43.036 14:05:47 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:43.036 14:05:47 -- common/autotest_common.sh@10 -- # set +x 00:06:43.036 ************************************ 00:06:43.036 START TEST app_cmdline 00:06:43.036 ************************************ 00:06:43.036 14:05:47 app_cmdline -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:43.036 * Looking for test storage... 00:06:43.036 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:43.036 14:05:47 app_cmdline -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:43.036 14:05:47 app_cmdline -- common/autotest_common.sh@1681 -- # lcov --version 00:06:43.036 14:05:47 app_cmdline -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:43.296 14:05:47 app_cmdline -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:06:43.296 14:05:47 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@345 -- # : 1 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:43.297 14:05:47 app_cmdline -- scripts/common.sh@368 -- # return 0 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:43.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.297 --rc genhtml_branch_coverage=1 00:06:43.297 --rc genhtml_function_coverage=1 00:06:43.297 --rc genhtml_legend=1 00:06:43.297 --rc geninfo_all_blocks=1 00:06:43.297 --rc geninfo_unexecuted_blocks=1 00:06:43.297 00:06:43.297 ' 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:43.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.297 --rc genhtml_branch_coverage=1 00:06:43.297 --rc genhtml_function_coverage=1 00:06:43.297 --rc genhtml_legend=1 00:06:43.297 --rc geninfo_all_blocks=1 00:06:43.297 --rc geninfo_unexecuted_blocks=1 00:06:43.297 00:06:43.297 ' 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:43.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.297 --rc genhtml_branch_coverage=1 00:06:43.297 --rc genhtml_function_coverage=1 00:06:43.297 --rc genhtml_legend=1 00:06:43.297 --rc geninfo_all_blocks=1 00:06:43.297 --rc geninfo_unexecuted_blocks=1 00:06:43.297 00:06:43.297 ' 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:43.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.297 --rc genhtml_branch_coverage=1 00:06:43.297 --rc genhtml_function_coverage=1 00:06:43.297 --rc genhtml_legend=1 00:06:43.297 --rc geninfo_all_blocks=1 00:06:43.297 --rc geninfo_unexecuted_blocks=1 00:06:43.297 00:06:43.297 ' 00:06:43.297 14:05:47 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:43.297 14:05:47 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=72474 00:06:43.297 14:05:47 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:43.297 14:05:47 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 72474 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 72474 ']' 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:43.297 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:43.297 14:05:47 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:43.297 [2024-09-30 14:05:47.839861] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:43.297 [2024-09-30 14:05:47.840045] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72474 ] 00:06:43.651 [2024-09-30 14:05:47.970458] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:43.651 [2024-09-30 14:05:47.998113] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:43.651 [2024-09-30 14:05:48.043614] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:06:44.221 { 00:06:44.221 "version": "SPDK v25.01-pre git sha1 09cc66129", 00:06:44.221 "fields": { 00:06:44.221 "major": 25, 00:06:44.221 "minor": 1, 00:06:44.221 "patch": 0, 00:06:44.221 "suffix": "-pre", 00:06:44.221 "commit": "09cc66129" 00:06:44.221 } 00:06:44.221 } 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@26 -- # sort 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:06:44.221 14:05:48 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:44.221 14:05:48 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:06:44.481 14:05:48 app_cmdline -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:44.481 request: 00:06:44.481 { 00:06:44.481 "method": "env_dpdk_get_mem_stats", 00:06:44.481 "req_id": 1 00:06:44.481 } 00:06:44.481 Got JSON-RPC error response 00:06:44.481 response: 00:06:44.481 { 00:06:44.481 "code": -32601, 00:06:44.481 "message": "Method not found" 00:06:44.481 } 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:44.481 14:05:49 app_cmdline -- app/cmdline.sh@1 -- # killprocess 72474 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 72474 ']' 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 72474 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72474 00:06:44.481 killing process with pid 72474 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:44.481 14:05:49 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:44.482 14:05:49 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72474' 00:06:44.482 14:05:49 app_cmdline -- common/autotest_common.sh@969 -- # kill 72474 00:06:44.482 14:05:49 app_cmdline -- common/autotest_common.sh@974 -- # wait 72474 00:06:45.050 00:06:45.050 real 0m1.961s 00:06:45.050 user 0m2.128s 00:06:45.050 sys 0m0.570s 00:06:45.050 14:05:49 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:45.050 14:05:49 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:45.050 ************************************ 00:06:45.050 END TEST app_cmdline 00:06:45.050 ************************************ 00:06:45.050 14:05:49 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:45.050 14:05:49 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:45.050 14:05:49 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:45.050 14:05:49 -- common/autotest_common.sh@10 -- # set +x 00:06:45.050 ************************************ 00:06:45.050 START TEST version 00:06:45.050 ************************************ 00:06:45.050 14:05:49 version -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:45.050 * Looking for test storage... 00:06:45.050 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:45.050 14:05:49 version -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:45.050 14:05:49 version -- common/autotest_common.sh@1681 -- # lcov --version 00:06:45.050 14:05:49 version -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:45.311 14:05:49 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:45.311 14:05:49 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:45.311 14:05:49 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:45.311 14:05:49 version -- scripts/common.sh@336 -- # IFS=.-: 00:06:45.311 14:05:49 version -- scripts/common.sh@336 -- # read -ra ver1 00:06:45.311 14:05:49 version -- scripts/common.sh@337 -- # IFS=.-: 00:06:45.311 14:05:49 version -- scripts/common.sh@337 -- # read -ra ver2 00:06:45.311 14:05:49 version -- scripts/common.sh@338 -- # local 'op=<' 00:06:45.311 14:05:49 version -- scripts/common.sh@340 -- # ver1_l=2 00:06:45.311 14:05:49 version -- scripts/common.sh@341 -- # ver2_l=1 00:06:45.311 14:05:49 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:45.311 14:05:49 version -- scripts/common.sh@344 -- # case "$op" in 00:06:45.311 14:05:49 version -- scripts/common.sh@345 -- # : 1 00:06:45.311 14:05:49 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:45.311 14:05:49 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:45.311 14:05:49 version -- scripts/common.sh@365 -- # decimal 1 00:06:45.311 14:05:49 version -- scripts/common.sh@353 -- # local d=1 00:06:45.311 14:05:49 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:45.311 14:05:49 version -- scripts/common.sh@355 -- # echo 1 00:06:45.311 14:05:49 version -- scripts/common.sh@365 -- # ver1[v]=1 00:06:45.311 14:05:49 version -- scripts/common.sh@366 -- # decimal 2 00:06:45.311 14:05:49 version -- scripts/common.sh@353 -- # local d=2 00:06:45.311 14:05:49 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:45.311 14:05:49 version -- scripts/common.sh@355 -- # echo 2 00:06:45.311 14:05:49 version -- scripts/common.sh@366 -- # ver2[v]=2 00:06:45.311 14:05:49 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:45.311 14:05:49 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:45.311 14:05:49 version -- scripts/common.sh@368 -- # return 0 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:45.311 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.311 --rc genhtml_branch_coverage=1 00:06:45.311 --rc genhtml_function_coverage=1 00:06:45.311 --rc genhtml_legend=1 00:06:45.311 --rc geninfo_all_blocks=1 00:06:45.311 --rc geninfo_unexecuted_blocks=1 00:06:45.311 00:06:45.311 ' 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:45.311 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.311 --rc genhtml_branch_coverage=1 00:06:45.311 --rc genhtml_function_coverage=1 00:06:45.311 --rc genhtml_legend=1 00:06:45.311 --rc geninfo_all_blocks=1 00:06:45.311 --rc geninfo_unexecuted_blocks=1 00:06:45.311 00:06:45.311 ' 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:45.311 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.311 --rc genhtml_branch_coverage=1 00:06:45.311 --rc genhtml_function_coverage=1 00:06:45.311 --rc genhtml_legend=1 00:06:45.311 --rc geninfo_all_blocks=1 00:06:45.311 --rc geninfo_unexecuted_blocks=1 00:06:45.311 00:06:45.311 ' 00:06:45.311 14:05:49 version -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:45.311 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.311 --rc genhtml_branch_coverage=1 00:06:45.311 --rc genhtml_function_coverage=1 00:06:45.311 --rc genhtml_legend=1 00:06:45.311 --rc geninfo_all_blocks=1 00:06:45.311 --rc geninfo_unexecuted_blocks=1 00:06:45.311 00:06:45.311 ' 00:06:45.311 14:05:49 version -- app/version.sh@17 -- # get_header_version major 00:06:45.311 14:05:49 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # tr -d '"' 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # cut -f2 00:06:45.311 14:05:49 version -- app/version.sh@17 -- # major=25 00:06:45.311 14:05:49 version -- app/version.sh@18 -- # get_header_version minor 00:06:45.311 14:05:49 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # cut -f2 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # tr -d '"' 00:06:45.311 14:05:49 version -- app/version.sh@18 -- # minor=1 00:06:45.311 14:05:49 version -- app/version.sh@19 -- # get_header_version patch 00:06:45.311 14:05:49 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # cut -f2 00:06:45.311 14:05:49 version -- app/version.sh@14 -- # tr -d '"' 00:06:45.312 14:05:49 version -- app/version.sh@19 -- # patch=0 00:06:45.312 14:05:49 version -- app/version.sh@20 -- # get_header_version suffix 00:06:45.312 14:05:49 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:45.312 14:05:49 version -- app/version.sh@14 -- # cut -f2 00:06:45.312 14:05:49 version -- app/version.sh@14 -- # tr -d '"' 00:06:45.312 14:05:49 version -- app/version.sh@20 -- # suffix=-pre 00:06:45.312 14:05:49 version -- app/version.sh@22 -- # version=25.1 00:06:45.312 14:05:49 version -- app/version.sh@25 -- # (( patch != 0 )) 00:06:45.312 14:05:49 version -- app/version.sh@28 -- # version=25.1rc0 00:06:45.312 14:05:49 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:06:45.312 14:05:49 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:06:45.312 14:05:49 version -- app/version.sh@30 -- # py_version=25.1rc0 00:06:45.312 14:05:49 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:06:45.312 ************************************ 00:06:45.312 END TEST version 00:06:45.312 ************************************ 00:06:45.312 00:06:45.312 real 0m0.326s 00:06:45.312 user 0m0.199s 00:06:45.312 sys 0m0.185s 00:06:45.312 14:05:49 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:45.312 14:05:49 version -- common/autotest_common.sh@10 -- # set +x 00:06:45.312 14:05:49 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:06:45.312 14:05:49 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:06:45.312 14:05:49 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:45.312 14:05:49 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:45.312 14:05:49 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:45.312 14:05:49 -- common/autotest_common.sh@10 -- # set +x 00:06:45.312 ************************************ 00:06:45.312 START TEST bdev_raid 00:06:45.312 ************************************ 00:06:45.312 14:05:49 bdev_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:45.572 * Looking for test storage... 00:06:45.572 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@345 -- # : 1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:45.572 14:05:50 bdev_raid -- scripts/common.sh@368 -- # return 0 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:45.572 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.572 --rc genhtml_branch_coverage=1 00:06:45.572 --rc genhtml_function_coverage=1 00:06:45.572 --rc genhtml_legend=1 00:06:45.572 --rc geninfo_all_blocks=1 00:06:45.572 --rc geninfo_unexecuted_blocks=1 00:06:45.572 00:06:45.572 ' 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:45.572 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.572 --rc genhtml_branch_coverage=1 00:06:45.572 --rc genhtml_function_coverage=1 00:06:45.572 --rc genhtml_legend=1 00:06:45.572 --rc geninfo_all_blocks=1 00:06:45.572 --rc geninfo_unexecuted_blocks=1 00:06:45.572 00:06:45.572 ' 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:45.572 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.572 --rc genhtml_branch_coverage=1 00:06:45.572 --rc genhtml_function_coverage=1 00:06:45.572 --rc genhtml_legend=1 00:06:45.572 --rc geninfo_all_blocks=1 00:06:45.572 --rc geninfo_unexecuted_blocks=1 00:06:45.572 00:06:45.572 ' 00:06:45.572 14:05:50 bdev_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:45.573 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:45.573 --rc genhtml_branch_coverage=1 00:06:45.573 --rc genhtml_function_coverage=1 00:06:45.573 --rc genhtml_legend=1 00:06:45.573 --rc geninfo_all_blocks=1 00:06:45.573 --rc geninfo_unexecuted_blocks=1 00:06:45.573 00:06:45.573 ' 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:45.573 14:05:50 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:06:45.573 14:05:50 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:06:45.573 14:05:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:45.573 14:05:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:45.573 14:05:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:45.573 ************************************ 00:06:45.573 START TEST raid1_resize_data_offset_test 00:06:45.573 ************************************ 00:06:45.573 Process raid pid: 72645 00:06:45.573 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1125 -- # raid_resize_data_offset_test 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=72645 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 72645' 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 72645 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@831 -- # '[' -z 72645 ']' 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:45.573 14:05:50 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.833 [2024-09-30 14:05:50.272757] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:45.833 [2024-09-30 14:05:50.272937] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:45.833 [2024-09-30 14:05:50.404621] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:45.833 [2024-09-30 14:05:50.434947] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:45.833 [2024-09-30 14:05:50.478973] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.093 [2024-09-30 14:05:50.520339] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:46.093 [2024-09-30 14:05:50.520463] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # return 0 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.662 malloc0 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.662 malloc1 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.662 null0 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.662 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.662 [2024-09-30 14:05:51.199614] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:06:46.662 [2024-09-30 14:05:51.201407] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:46.662 [2024-09-30 14:05:51.201516] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:06:46.662 [2024-09-30 14:05:51.201700] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:46.663 [2024-09-30 14:05:51.201763] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:06:46.663 [2024-09-30 14:05:51.202066] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:06:46.663 [2024-09-30 14:05:51.202261] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:46.663 [2024-09-30 14:05:51.202304] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:06:46.663 [2024-09-30 14:05:51.202515] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.663 [2024-09-30 14:05:51.255447] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.663 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.923 malloc2 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.923 [2024-09-30 14:05:51.380584] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:46.923 [2024-09-30 14:05:51.384862] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.923 [2024-09-30 14:05:51.386726] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 72645 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@950 -- # '[' -z 72645 ']' 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # kill -0 72645 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # uname 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72645 00:06:46.923 killing process with pid 72645 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72645' 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@969 -- # kill 72645 00:06:46.923 [2024-09-30 14:05:51.485690] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:46.923 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@974 -- # wait 72645 00:06:46.923 [2024-09-30 14:05:51.486971] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:06:46.923 [2024-09-30 14:05:51.487031] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:46.923 [2024-09-30 14:05:51.487049] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:06:46.923 [2024-09-30 14:05:51.491936] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:46.923 [2024-09-30 14:05:51.492207] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:46.923 [2024-09-30 14:05:51.492219] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:06:47.183 [2024-09-30 14:05:51.697786] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:47.443 14:05:51 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:06:47.443 00:06:47.443 real 0m1.753s 00:06:47.443 user 0m1.699s 00:06:47.443 sys 0m0.499s 00:06:47.443 ************************************ 00:06:47.443 END TEST raid1_resize_data_offset_test 00:06:47.443 ************************************ 00:06:47.443 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:47.443 14:05:51 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.443 14:05:52 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:06:47.443 14:05:52 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:47.443 14:05:52 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:47.443 14:05:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:47.443 ************************************ 00:06:47.443 START TEST raid0_resize_superblock_test 00:06:47.443 ************************************ 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 0 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=72690 00:06:47.443 Process raid pid: 72690 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 72690' 00:06:47.443 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 72690 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 72690 ']' 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:47.443 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.443 [2024-09-30 14:05:52.096846] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:47.443 [2024-09-30 14:05:52.096961] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:47.703 [2024-09-30 14:05:52.228749] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:47.703 [2024-09-30 14:05:52.256377] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:47.703 [2024-09-30 14:05:52.303535] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.703 [2024-09-30 14:05:52.345069] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:47.703 [2024-09-30 14:05:52.345106] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:48.273 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:48.273 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:06:48.273 14:05:52 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:48.273 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.273 14:05:52 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.532 malloc0 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.532 [2024-09-30 14:05:53.053737] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:48.532 [2024-09-30 14:05:53.053850] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:48.532 [2024-09-30 14:05:53.053877] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:06:48.532 [2024-09-30 14:05:53.053889] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:48.532 [2024-09-30 14:05:53.055993] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:48.532 [2024-09-30 14:05:53.056037] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:48.532 pt0 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.532 1b5825c7-dacb-4a5f-a2cf-9edbc85e5c44 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.532 e7a66c8b-7310-436c-8de9-8ab225d1e389 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.532 38bbfe98-a847-4d6c-a41d-a914353ee12d 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.532 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 [2024-09-30 14:05:53.187955] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev e7a66c8b-7310-436c-8de9-8ab225d1e389 is claimed 00:06:48.793 [2024-09-30 14:05:53.188045] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 38bbfe98-a847-4d6c-a41d-a914353ee12d is claimed 00:06:48.793 [2024-09-30 14:05:53.188158] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:48.793 [2024-09-30 14:05:53.188168] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:06:48.793 [2024-09-30 14:05:53.188418] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:06:48.793 [2024-09-30 14:05:53.188583] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:48.793 [2024-09-30 14:05:53.188596] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:06:48.793 [2024-09-30 14:05:53.188752] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 [2024-09-30 14:05:53.303934] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 [2024-09-30 14:05:53.347813] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:48.793 [2024-09-30 14:05:53.347841] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'e7a66c8b-7310-436c-8de9-8ab225d1e389' was resized: old size 131072, new size 204800 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 [2024-09-30 14:05:53.359706] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:48.793 [2024-09-30 14:05:53.359731] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '38bbfe98-a847-4d6c-a41d-a914353ee12d' was resized: old size 131072, new size 204800 00:06:48.793 [2024-09-30 14:05:53.359755] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.793 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.054 [2024-09-30 14:05:53.467644] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.054 [2024-09-30 14:05:53.511343] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:49.054 [2024-09-30 14:05:53.511447] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:49.054 [2024-09-30 14:05:53.511489] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:49.054 [2024-09-30 14:05:53.511530] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:49.054 [2024-09-30 14:05:53.511646] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:49.054 [2024-09-30 14:05:53.511719] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:49.054 [2024-09-30 14:05:53.511765] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.054 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.054 [2024-09-30 14:05:53.519316] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:49.054 [2024-09-30 14:05:53.519400] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:49.054 [2024-09-30 14:05:53.519456] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:06:49.054 [2024-09-30 14:05:53.519500] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:49.054 [2024-09-30 14:05:53.521565] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:49.054 [2024-09-30 14:05:53.521627] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:49.054 [2024-09-30 14:05:53.523018] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev e7a66c8b-7310-436c-8de9-8ab225d1e389 00:06:49.054 [2024-09-30 14:05:53.523113] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev e7a66c8b-7310-436c-8de9-8ab225d1e389 is claimed 00:06:49.054 [2024-09-30 14:05:53.523234] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 38bbfe98-a847-4d6c-a41d-a914353ee12d 00:06:49.054 [2024-09-30 14:05:53.523293] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 38bbfe98-a847-4d6c-a41d-a914353ee12d is claimed 00:06:49.055 [2024-09-30 14:05:53.523424] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 38bbfe98-a847-4d6c-a41d-a914353ee12d (2) smaller than existing raid bdev Raid (3) 00:06:49.055 [2024-09-30 14:05:53.523491] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev e7a66c8b-7310-436c-8de9-8ab225d1e389: File exists 00:06:49.055 [2024-09-30 14:05:53.523583] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:06:49.055 [2024-09-30 14:05:53.523614] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:06:49.055 [2024-09-30 14:05:53.523845] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:06:49.055 [2024-09-30 14:05:53.523979] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:06:49.055 [2024-09-30 14:05:53.524017] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:06:49.055 [2024-09-30 14:05:53.524178] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:49.055 pt0 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:06:49.055 [2024-09-30 14:05:53.543675] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 72690 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 72690 ']' 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 72690 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72690 00:06:49.055 killing process with pid 72690 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72690' 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 72690 00:06:49.055 [2024-09-30 14:05:53.630880] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:49.055 [2024-09-30 14:05:53.630935] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:49.055 [2024-09-30 14:05:53.630966] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:49.055 [2024-09-30 14:05:53.630976] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:06:49.055 14:05:53 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 72690 00:06:49.315 [2024-09-30 14:05:53.786410] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:49.574 ************************************ 00:06:49.574 END TEST raid0_resize_superblock_test 00:06:49.574 ************************************ 00:06:49.574 14:05:54 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:49.574 00:06:49.574 real 0m2.015s 00:06:49.574 user 0m2.289s 00:06:49.574 sys 0m0.527s 00:06:49.574 14:05:54 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:49.574 14:05:54 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.574 14:05:54 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:06:49.574 14:05:54 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:49.574 14:05:54 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:49.574 14:05:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:49.574 ************************************ 00:06:49.574 START TEST raid1_resize_superblock_test 00:06:49.574 ************************************ 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 1 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=72761 00:06:49.574 Process raid pid: 72761 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 72761' 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 72761 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 72761 ']' 00:06:49.574 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:49.574 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.574 [2024-09-30 14:05:54.184149] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:49.574 [2024-09-30 14:05:54.184356] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:49.834 [2024-09-30 14:05:54.315676] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:49.834 [2024-09-30 14:05:54.330379] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:49.834 [2024-09-30 14:05:54.375099] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:49.834 [2024-09-30 14:05:54.416747] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:49.834 [2024-09-30 14:05:54.416849] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:50.403 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:50.403 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:06:50.403 14:05:54 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:50.403 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.403 14:05:54 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 malloc0 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 [2024-09-30 14:05:55.129179] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:50.664 [2024-09-30 14:05:55.129297] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:50.664 [2024-09-30 14:05:55.129342] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:06:50.664 [2024-09-30 14:05:55.129375] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:50.664 [2024-09-30 14:05:55.131401] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:50.664 [2024-09-30 14:05:55.131473] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:50.664 pt0 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 efc1dfa2-6bfb-4077-a47e-2629e00f2a00 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 ecb17dfb-9e47-4e04-ae03-398d10af1f7d 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 [2024-09-30 14:05:55.262667] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev ecb17dfb-9e47-4e04-ae03-398d10af1f7d is claimed 00:06:50.664 [2024-09-30 14:05:55.262745] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b is claimed 00:06:50.664 [2024-09-30 14:05:55.262871] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:50.664 [2024-09-30 14:05:55.262885] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:06:50.664 [2024-09-30 14:05:55.263129] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:06:50.664 [2024-09-30 14:05:55.263298] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:50.664 [2024-09-30 14:05:55.263313] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:06:50.664 [2024-09-30 14:05:55.263436] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.664 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:06:50.924 [2024-09-30 14:05:55.374721] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 [2024-09-30 14:05:55.422536] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:50.924 [2024-09-30 14:05:55.422601] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'ecb17dfb-9e47-4e04-ae03-398d10af1f7d' was resized: old size 131072, new size 204800 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 [2024-09-30 14:05:55.434482] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:50.924 [2024-09-30 14:05:55.434518] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b' was resized: old size 131072, new size 204800 00:06:50.924 [2024-09-30 14:05:55.434549] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:06:50.924 [2024-09-30 14:05:55.546367] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:50.924 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.184 [2024-09-30 14:05:55.590099] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:51.184 [2024-09-30 14:05:55.590163] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:51.184 [2024-09-30 14:05:55.590190] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:51.184 [2024-09-30 14:05:55.590337] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:51.184 [2024-09-30 14:05:55.590515] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:51.184 [2024-09-30 14:05:55.590571] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:51.184 [2024-09-30 14:05:55.590582] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.184 [2024-09-30 14:05:55.602055] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:51.184 [2024-09-30 14:05:55.602100] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:51.184 [2024-09-30 14:05:55.602122] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:06:51.184 [2024-09-30 14:05:55.602130] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:51.184 [2024-09-30 14:05:55.604094] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:51.184 [2024-09-30 14:05:55.604126] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:51.184 [2024-09-30 14:05:55.605428] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev ecb17dfb-9e47-4e04-ae03-398d10af1f7d 00:06:51.184 [2024-09-30 14:05:55.605548] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev ecb17dfb-9e47-4e04-ae03-398d10af1f7d is claimed 00:06:51.184 [2024-09-30 14:05:55.605635] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b 00:06:51.184 [2024-09-30 14:05:55.605650] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b is claimed 00:06:51.184 [2024-09-30 14:05:55.605735] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev f5f5d8b6-c5ae-49ee-a3ac-a4053d04ab8b (2) smaller than existing raid bdev Raid (3) 00:06:51.184 [2024-09-30 14:05:55.605756] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev ecb17dfb-9e47-4e04-ae03-398d10af1f7d: File exists 00:06:51.184 [2024-09-30 14:05:55.605811] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:06:51.184 [2024-09-30 14:05:55.605817] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:06:51.184 [2024-09-30 14:05:55.606020] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:06:51.184 [2024-09-30 14:05:55.606128] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:06:51.184 [2024-09-30 14:05:55.606138] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:06:51.184 [2024-09-30 14:05:55.606261] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:51.184 pt0 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:51.184 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.184 [2024-09-30 14:05:55.630412] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 72761 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 72761 ']' 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 72761 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72761 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:51.185 killing process with pid 72761 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72761' 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 72761 00:06:51.185 [2024-09-30 14:05:55.711312] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:51.185 [2024-09-30 14:05:55.711377] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:51.185 [2024-09-30 14:05:55.711417] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:51.185 [2024-09-30 14:05:55.711427] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:06:51.185 14:05:55 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 72761 00:06:51.444 [2024-09-30 14:05:55.868855] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:51.703 14:05:56 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:51.703 00:06:51.703 real 0m2.010s 00:06:51.703 user 0m2.259s 00:06:51.703 sys 0m0.481s 00:06:51.703 14:05:56 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:51.703 ************************************ 00:06:51.703 END TEST raid1_resize_superblock_test 00:06:51.703 ************************************ 00:06:51.703 14:05:56 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:06:51.703 14:05:56 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:06:51.703 14:05:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:51.703 14:05:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:51.703 14:05:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:51.703 ************************************ 00:06:51.703 START TEST raid_function_test_raid0 00:06:51.703 ************************************ 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1125 -- # raid_function_test raid0 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=72839 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 72839' 00:06:51.703 Process raid pid: 72839 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 72839 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@831 -- # '[' -z 72839 ']' 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:51.703 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:51.703 14:05:56 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:51.703 [2024-09-30 14:05:56.296800] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:51.703 [2024-09-30 14:05:56.297043] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:51.962 [2024-09-30 14:05:56.437458] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:51.962 [2024-09-30 14:05:56.465149] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:51.962 [2024-09-30 14:05:56.511296] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:51.962 [2024-09-30 14:05:56.552862] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:51.962 [2024-09-30 14:05:56.552902] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # return 0 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:52.530 Base_1 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:52.530 Base_2 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:52.530 [2024-09-30 14:05:57.146844] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:52.530 [2024-09-30 14:05:57.148570] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:52.530 [2024-09-30 14:05:57.148632] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:52.530 [2024-09-30 14:05:57.148642] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:52.530 [2024-09-30 14:05:57.148894] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:06:52.530 [2024-09-30 14:05:57.149004] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:52.530 [2024-09-30 14:05:57.149033] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:06:52.530 [2024-09-30 14:05:57.149159] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:52.530 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:52.789 [2024-09-30 14:05:57.382550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:06:52.789 /dev/nbd0 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # local i 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # break 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:52.789 1+0 records in 00:06:52.789 1+0 records out 00:06:52.789 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000333209 s, 12.3 MB/s 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # size=4096 00:06:52.789 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # return 0 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:53.048 { 00:06:53.048 "nbd_device": "/dev/nbd0", 00:06:53.048 "bdev_name": "raid" 00:06:53.048 } 00:06:53.048 ]' 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:53.048 { 00:06:53.048 "nbd_device": "/dev/nbd0", 00:06:53.048 "bdev_name": "raid" 00:06:53.048 } 00:06:53.048 ]' 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:06:53.048 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:53.307 4096+0 records in 00:06:53.307 4096+0 records out 00:06:53.307 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0324793 s, 64.6 MB/s 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:53.307 4096+0 records in 00:06:53.307 4096+0 records out 00:06:53.307 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.174495 s, 12.0 MB/s 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:53.307 128+0 records in 00:06:53.307 128+0 records out 00:06:53.307 65536 bytes (66 kB, 64 KiB) copied, 0.00122356 s, 53.6 MB/s 00:06:53.307 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:53.567 2035+0 records in 00:06:53.567 2035+0 records out 00:06:53.567 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0144818 s, 71.9 MB/s 00:06:53.567 14:05:57 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:53.567 456+0 records in 00:06:53.567 456+0 records out 00:06:53.567 233472 bytes (233 kB, 228 KiB) copied, 0.00346766 s, 67.3 MB/s 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:53.567 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:53.827 [2024-09-30 14:05:58.250319] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:53.827 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:53.828 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:53.828 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 72839 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@950 -- # '[' -z 72839 ']' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # kill -0 72839 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # uname 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72839 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:54.087 killing process with pid 72839 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72839' 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@969 -- # kill 72839 00:06:54.087 [2024-09-30 14:05:58.565187] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:54.087 [2024-09-30 14:05:58.565288] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:54.087 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@974 -- # wait 72839 00:06:54.087 [2024-09-30 14:05:58.565347] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:54.087 [2024-09-30 14:05:58.565363] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:06:54.087 [2024-09-30 14:05:58.587542] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:54.347 14:05:58 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:06:54.347 00:06:54.347 real 0m2.631s 00:06:54.347 user 0m3.204s 00:06:54.347 sys 0m0.927s 00:06:54.347 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:54.347 14:05:58 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:54.347 ************************************ 00:06:54.347 END TEST raid_function_test_raid0 00:06:54.347 ************************************ 00:06:54.347 14:05:58 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:06:54.347 14:05:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:54.347 14:05:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:54.347 14:05:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:54.347 ************************************ 00:06:54.347 START TEST raid_function_test_concat 00:06:54.347 ************************************ 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1125 -- # raid_function_test concat 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=72957 00:06:54.347 Process raid pid: 72957 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 72957' 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 72957 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@831 -- # '[' -z 72957 ']' 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:54.347 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:54.347 14:05:58 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:54.347 [2024-09-30 14:05:58.990812] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:54.347 [2024-09-30 14:05:58.990948] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:54.606 [2024-09-30 14:05:59.123124] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:54.606 [2024-09-30 14:05:59.152064] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:54.606 [2024-09-30 14:05:59.203414] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:54.606 [2024-09-30 14:05:59.244591] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:54.606 [2024-09-30 14:05:59.244624] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:55.174 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:55.174 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # return 0 00:06:55.174 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:55.174 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:55.174 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:55.433 Base_1 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:55.433 Base_2 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:55.433 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:55.433 [2024-09-30 14:05:59.867020] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:55.433 [2024-09-30 14:05:59.868793] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:55.433 [2024-09-30 14:05:59.868870] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:55.433 [2024-09-30 14:05:59.868891] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:55.433 [2024-09-30 14:05:59.869169] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:06:55.433 [2024-09-30 14:05:59.869299] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:55.434 [2024-09-30 14:05:59.869312] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:06:55.434 [2024-09-30 14:05:59.869444] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:55.434 14:05:59 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:55.434 [2024-09-30 14:06:00.066756] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:06:55.434 /dev/nbd0 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # local i 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # break 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:55.693 1+0 records in 00:06:55.693 1+0 records out 00:06:55.693 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000193964 s, 21.1 MB/s 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # size=4096 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # return 0 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:55.693 { 00:06:55.693 "nbd_device": "/dev/nbd0", 00:06:55.693 "bdev_name": "raid" 00:06:55.693 } 00:06:55.693 ]' 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:55.693 { 00:06:55.693 "nbd_device": "/dev/nbd0", 00:06:55.693 "bdev_name": "raid" 00:06:55.693 } 00:06:55.693 ]' 00:06:55.693 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:55.951 4096+0 records in 00:06:55.951 4096+0 records out 00:06:55.951 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0362105 s, 57.9 MB/s 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:55.951 4096+0 records in 00:06:55.951 4096+0 records out 00:06:55.951 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.179485 s, 11.7 MB/s 00:06:55.951 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:56.212 128+0 records in 00:06:56.212 128+0 records out 00:06:56.212 65536 bytes (66 kB, 64 KiB) copied, 0.000356809 s, 184 MB/s 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:56.212 2035+0 records in 00:06:56.212 2035+0 records out 00:06:56.212 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0149004 s, 69.9 MB/s 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:56.212 456+0 records in 00:06:56.212 456+0 records out 00:06:56.212 233472 bytes (233 kB, 228 KiB) copied, 0.00363877 s, 64.2 MB/s 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:56.212 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:56.472 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:56.473 [2024-09-30 14:06:00.928944] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:56.473 14:06:00 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 72957 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@950 -- # '[' -z 72957 ']' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # kill -0 72957 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # uname 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72957 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:56.732 killing process with pid 72957 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72957' 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@969 -- # kill 72957 00:06:56.732 [2024-09-30 14:06:01.234461] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:56.732 [2024-09-30 14:06:01.234585] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:56.732 [2024-09-30 14:06:01.234647] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:56.732 [2024-09-30 14:06:01.234657] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:06:56.732 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@974 -- # wait 72957 00:06:56.732 [2024-09-30 14:06:01.256451] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:56.989 14:06:01 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:06:56.989 00:06:56.989 real 0m2.592s 00:06:56.989 user 0m3.165s 00:06:56.989 sys 0m0.896s 00:06:56.989 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:56.989 14:06:01 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:56.989 ************************************ 00:06:56.989 END TEST raid_function_test_concat 00:06:56.989 ************************************ 00:06:56.989 14:06:01 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:06:56.989 14:06:01 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:56.989 14:06:01 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:56.989 14:06:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:56.989 ************************************ 00:06:56.989 START TEST raid0_resize_test 00:06:56.989 ************************************ 00:06:56.989 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 0 00:06:56.989 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:06:56.989 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=73069 00:06:56.990 Process raid pid: 73069 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 73069' 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 73069 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@831 -- # '[' -z 73069 ']' 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:56.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:56.990 14:06:01 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.248 [2024-09-30 14:06:01.653891] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:57.248 [2024-09-30 14:06:01.654012] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:57.248 [2024-09-30 14:06:01.787799] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:57.248 [2024-09-30 14:06:01.817116] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:57.248 [2024-09-30 14:06:01.862681] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:57.507 [2024-09-30 14:06:01.904751] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:57.507 [2024-09-30 14:06:01.904785] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # return 0 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.078 Base_1 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.078 Base_2 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.078 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.078 [2024-09-30 14:06:02.506172] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:58.078 [2024-09-30 14:06:02.508067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:58.078 [2024-09-30 14:06:02.508133] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:58.078 [2024-09-30 14:06:02.508142] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:58.078 [2024-09-30 14:06:02.508436] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:06:58.079 [2024-09-30 14:06:02.508559] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:58.079 [2024-09-30 14:06:02.508572] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:06:58.079 [2024-09-30 14:06:02.508719] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.079 [2024-09-30 14:06:02.518104] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:58.079 [2024-09-30 14:06:02.518129] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:58.079 true 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.079 [2024-09-30 14:06:02.534281] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.079 [2024-09-30 14:06:02.578042] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:58.079 [2024-09-30 14:06:02.578080] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:58.079 [2024-09-30 14:06:02.578111] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:06:58.079 true 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:58.079 [2024-09-30 14:06:02.590182] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 73069 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@950 -- # '[' -z 73069 ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # kill -0 73069 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # uname 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73069 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:58.079 killing process with pid 73069 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73069' 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@969 -- # kill 73069 00:06:58.079 [2024-09-30 14:06:02.674971] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:58.079 [2024-09-30 14:06:02.675081] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:58.079 [2024-09-30 14:06:02.675130] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:58.079 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@974 -- # wait 73069 00:06:58.079 [2024-09-30 14:06:02.675148] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:06:58.079 [2024-09-30 14:06:02.676642] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:58.371 14:06:02 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:58.371 00:06:58.371 real 0m1.352s 00:06:58.371 user 0m1.497s 00:06:58.371 sys 0m0.316s 00:06:58.371 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:58.371 14:06:02 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.371 ************************************ 00:06:58.371 END TEST raid0_resize_test 00:06:58.371 ************************************ 00:06:58.371 14:06:02 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:06:58.371 14:06:02 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:58.371 14:06:02 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:58.371 14:06:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:58.371 ************************************ 00:06:58.371 START TEST raid1_resize_test 00:06:58.371 ************************************ 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 1 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=73119 00:06:58.371 Process raid pid: 73119 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 73119' 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 73119 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@831 -- # '[' -z 73119 ']' 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:58.371 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:58.371 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:58.372 14:06:02 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.630 [2024-09-30 14:06:03.081176] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:58.630 [2024-09-30 14:06:03.081335] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:58.630 [2024-09-30 14:06:03.219220] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:06:58.630 [2024-09-30 14:06:03.248562] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:58.888 [2024-09-30 14:06:03.297640] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:58.888 [2024-09-30 14:06:03.339380] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:58.888 [2024-09-30 14:06:03.339419] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # return 0 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.457 Base_1 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.457 Base_2 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.457 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.458 [2024-09-30 14:06:03.920418] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:59.458 [2024-09-30 14:06:03.922218] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:59.458 [2024-09-30 14:06:03.922283] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:06:59.458 [2024-09-30 14:06:03.922294] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:06:59.458 [2024-09-30 14:06:03.922578] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:06:59.458 [2024-09-30 14:06:03.922694] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:06:59.458 [2024-09-30 14:06:03.922706] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:06:59.458 [2024-09-30 14:06:03.922828] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.458 [2024-09-30 14:06:03.932381] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:59.458 [2024-09-30 14:06:03.932409] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:59.458 true 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.458 [2024-09-30 14:06:03.948528] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.458 [2024-09-30 14:06:03.992264] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:59.458 [2024-09-30 14:06:03.992292] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:59.458 [2024-09-30 14:06:03.992318] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:06:59.458 true 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:59.458 14:06:03 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.458 [2024-09-30 14:06:04.008418] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 73119 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@950 -- # '[' -z 73119 ']' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # kill -0 73119 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # uname 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73119 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:59.458 killing process with pid 73119 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73119' 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@969 -- # kill 73119 00:06:59.458 [2024-09-30 14:06:04.086611] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:59.458 [2024-09-30 14:06:04.086706] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:59.458 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@974 -- # wait 73119 00:06:59.458 [2024-09-30 14:06:04.087167] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:59.458 [2024-09-30 14:06:04.087190] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:06:59.458 [2024-09-30 14:06:04.088284] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:59.717 14:06:04 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:59.717 00:06:59.717 real 0m1.345s 00:06:59.717 user 0m1.483s 00:06:59.717 sys 0m0.328s 00:06:59.717 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:59.717 14:06:04 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.717 ************************************ 00:06:59.717 END TEST raid1_resize_test 00:06:59.717 ************************************ 00:06:59.977 14:06:04 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:06:59.977 14:06:04 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:59.977 14:06:04 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:06:59.977 14:06:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:06:59.977 14:06:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:59.977 14:06:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:59.977 ************************************ 00:06:59.977 START TEST raid_state_function_test 00:06:59.977 ************************************ 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 false 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73165 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73165' 00:06:59.977 Process raid pid: 73165 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73165 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 73165 ']' 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:59.977 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:59.977 14:06:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.977 [2024-09-30 14:06:04.491605] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:06:59.977 [2024-09-30 14:06:04.491756] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:59.977 [2024-09-30 14:06:04.623268] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:00.237 [2024-09-30 14:06:04.651982] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:00.237 [2024-09-30 14:06:04.696639] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:00.237 [2024-09-30 14:06:04.737574] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:00.237 [2024-09-30 14:06:04.737607] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.804 [2024-09-30 14:06:05.318625] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:00.804 [2024-09-30 14:06:05.318686] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:00.804 [2024-09-30 14:06:05.318699] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.804 [2024-09-30 14:06:05.318723] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.804 "name": "Existed_Raid", 00:07:00.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.804 "strip_size_kb": 64, 00:07:00.804 "state": "configuring", 00:07:00.804 "raid_level": "raid0", 00:07:00.804 "superblock": false, 00:07:00.804 "num_base_bdevs": 2, 00:07:00.804 "num_base_bdevs_discovered": 0, 00:07:00.804 "num_base_bdevs_operational": 2, 00:07:00.804 "base_bdevs_list": [ 00:07:00.804 { 00:07:00.804 "name": "BaseBdev1", 00:07:00.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.804 "is_configured": false, 00:07:00.804 "data_offset": 0, 00:07:00.804 "data_size": 0 00:07:00.804 }, 00:07:00.804 { 00:07:00.804 "name": "BaseBdev2", 00:07:00.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.804 "is_configured": false, 00:07:00.804 "data_offset": 0, 00:07:00.804 "data_size": 0 00:07:00.804 } 00:07:00.804 ] 00:07:00.804 }' 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.804 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.373 [2024-09-30 14:06:05.733806] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:01.373 [2024-09-30 14:06:05.733844] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.373 [2024-09-30 14:06:05.745764] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:01.373 [2024-09-30 14:06:05.745802] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:01.373 [2024-09-30 14:06:05.745811] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:01.373 [2024-09-30 14:06:05.745818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.373 [2024-09-30 14:06:05.766223] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:01.373 BaseBdev1 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.373 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.374 [ 00:07:01.374 { 00:07:01.374 "name": "BaseBdev1", 00:07:01.374 "aliases": [ 00:07:01.374 "f13be733-dbe5-4806-9f39-12072cd679ac" 00:07:01.374 ], 00:07:01.374 "product_name": "Malloc disk", 00:07:01.374 "block_size": 512, 00:07:01.374 "num_blocks": 65536, 00:07:01.374 "uuid": "f13be733-dbe5-4806-9f39-12072cd679ac", 00:07:01.374 "assigned_rate_limits": { 00:07:01.374 "rw_ios_per_sec": 0, 00:07:01.374 "rw_mbytes_per_sec": 0, 00:07:01.374 "r_mbytes_per_sec": 0, 00:07:01.374 "w_mbytes_per_sec": 0 00:07:01.374 }, 00:07:01.374 "claimed": true, 00:07:01.374 "claim_type": "exclusive_write", 00:07:01.374 "zoned": false, 00:07:01.374 "supported_io_types": { 00:07:01.374 "read": true, 00:07:01.374 "write": true, 00:07:01.374 "unmap": true, 00:07:01.374 "flush": true, 00:07:01.374 "reset": true, 00:07:01.374 "nvme_admin": false, 00:07:01.374 "nvme_io": false, 00:07:01.374 "nvme_io_md": false, 00:07:01.374 "write_zeroes": true, 00:07:01.374 "zcopy": true, 00:07:01.374 "get_zone_info": false, 00:07:01.374 "zone_management": false, 00:07:01.374 "zone_append": false, 00:07:01.374 "compare": false, 00:07:01.374 "compare_and_write": false, 00:07:01.374 "abort": true, 00:07:01.374 "seek_hole": false, 00:07:01.374 "seek_data": false, 00:07:01.374 "copy": true, 00:07:01.374 "nvme_iov_md": false 00:07:01.374 }, 00:07:01.374 "memory_domains": [ 00:07:01.374 { 00:07:01.374 "dma_device_id": "system", 00:07:01.374 "dma_device_type": 1 00:07:01.374 }, 00:07:01.374 { 00:07:01.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:01.374 "dma_device_type": 2 00:07:01.374 } 00:07:01.374 ], 00:07:01.374 "driver_specific": {} 00:07:01.374 } 00:07:01.374 ] 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.374 "name": "Existed_Raid", 00:07:01.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.374 "strip_size_kb": 64, 00:07:01.374 "state": "configuring", 00:07:01.374 "raid_level": "raid0", 00:07:01.374 "superblock": false, 00:07:01.374 "num_base_bdevs": 2, 00:07:01.374 "num_base_bdevs_discovered": 1, 00:07:01.374 "num_base_bdevs_operational": 2, 00:07:01.374 "base_bdevs_list": [ 00:07:01.374 { 00:07:01.374 "name": "BaseBdev1", 00:07:01.374 "uuid": "f13be733-dbe5-4806-9f39-12072cd679ac", 00:07:01.374 "is_configured": true, 00:07:01.374 "data_offset": 0, 00:07:01.374 "data_size": 65536 00:07:01.374 }, 00:07:01.374 { 00:07:01.374 "name": "BaseBdev2", 00:07:01.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.374 "is_configured": false, 00:07:01.374 "data_offset": 0, 00:07:01.374 "data_size": 0 00:07:01.374 } 00:07:01.374 ] 00:07:01.374 }' 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.374 14:06:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.633 [2024-09-30 14:06:06.189572] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:01.633 [2024-09-30 14:06:06.189632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.633 [2024-09-30 14:06:06.201545] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:01.633 [2024-09-30 14:06:06.203325] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:01.633 [2024-09-30 14:06:06.203365] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.633 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.633 "name": "Existed_Raid", 00:07:01.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.633 "strip_size_kb": 64, 00:07:01.633 "state": "configuring", 00:07:01.633 "raid_level": "raid0", 00:07:01.633 "superblock": false, 00:07:01.633 "num_base_bdevs": 2, 00:07:01.634 "num_base_bdevs_discovered": 1, 00:07:01.634 "num_base_bdevs_operational": 2, 00:07:01.634 "base_bdevs_list": [ 00:07:01.634 { 00:07:01.634 "name": "BaseBdev1", 00:07:01.634 "uuid": "f13be733-dbe5-4806-9f39-12072cd679ac", 00:07:01.634 "is_configured": true, 00:07:01.634 "data_offset": 0, 00:07:01.634 "data_size": 65536 00:07:01.634 }, 00:07:01.634 { 00:07:01.634 "name": "BaseBdev2", 00:07:01.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.634 "is_configured": false, 00:07:01.634 "data_offset": 0, 00:07:01.634 "data_size": 0 00:07:01.634 } 00:07:01.634 ] 00:07:01.634 }' 00:07:01.634 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.634 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.200 [2024-09-30 14:06:06.641046] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:02.200 [2024-09-30 14:06:06.641089] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:02.200 [2024-09-30 14:06:06.641101] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:02.200 [2024-09-30 14:06:06.641428] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:02.200 [2024-09-30 14:06:06.641617] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:02.200 [2024-09-30 14:06:06.641637] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:02.200 [2024-09-30 14:06:06.641853] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:02.200 BaseBdev2 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.200 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.201 [ 00:07:02.201 { 00:07:02.201 "name": "BaseBdev2", 00:07:02.201 "aliases": [ 00:07:02.201 "93db4cea-47aa-455a-a128-165da801217e" 00:07:02.201 ], 00:07:02.201 "product_name": "Malloc disk", 00:07:02.201 "block_size": 512, 00:07:02.201 "num_blocks": 65536, 00:07:02.201 "uuid": "93db4cea-47aa-455a-a128-165da801217e", 00:07:02.201 "assigned_rate_limits": { 00:07:02.201 "rw_ios_per_sec": 0, 00:07:02.201 "rw_mbytes_per_sec": 0, 00:07:02.201 "r_mbytes_per_sec": 0, 00:07:02.201 "w_mbytes_per_sec": 0 00:07:02.201 }, 00:07:02.201 "claimed": true, 00:07:02.201 "claim_type": "exclusive_write", 00:07:02.201 "zoned": false, 00:07:02.201 "supported_io_types": { 00:07:02.201 "read": true, 00:07:02.201 "write": true, 00:07:02.201 "unmap": true, 00:07:02.201 "flush": true, 00:07:02.201 "reset": true, 00:07:02.201 "nvme_admin": false, 00:07:02.201 "nvme_io": false, 00:07:02.201 "nvme_io_md": false, 00:07:02.201 "write_zeroes": true, 00:07:02.201 "zcopy": true, 00:07:02.201 "get_zone_info": false, 00:07:02.201 "zone_management": false, 00:07:02.201 "zone_append": false, 00:07:02.201 "compare": false, 00:07:02.201 "compare_and_write": false, 00:07:02.201 "abort": true, 00:07:02.201 "seek_hole": false, 00:07:02.201 "seek_data": false, 00:07:02.201 "copy": true, 00:07:02.201 "nvme_iov_md": false 00:07:02.201 }, 00:07:02.201 "memory_domains": [ 00:07:02.201 { 00:07:02.201 "dma_device_id": "system", 00:07:02.201 "dma_device_type": 1 00:07:02.201 }, 00:07:02.201 { 00:07:02.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:02.201 "dma_device_type": 2 00:07:02.201 } 00:07:02.201 ], 00:07:02.201 "driver_specific": {} 00:07:02.201 } 00:07:02.201 ] 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:02.201 "name": "Existed_Raid", 00:07:02.201 "uuid": "9d3bae51-f7e9-4f8f-8eb4-732b6ecf20dd", 00:07:02.201 "strip_size_kb": 64, 00:07:02.201 "state": "online", 00:07:02.201 "raid_level": "raid0", 00:07:02.201 "superblock": false, 00:07:02.201 "num_base_bdevs": 2, 00:07:02.201 "num_base_bdevs_discovered": 2, 00:07:02.201 "num_base_bdevs_operational": 2, 00:07:02.201 "base_bdevs_list": [ 00:07:02.201 { 00:07:02.201 "name": "BaseBdev1", 00:07:02.201 "uuid": "f13be733-dbe5-4806-9f39-12072cd679ac", 00:07:02.201 "is_configured": true, 00:07:02.201 "data_offset": 0, 00:07:02.201 "data_size": 65536 00:07:02.201 }, 00:07:02.201 { 00:07:02.201 "name": "BaseBdev2", 00:07:02.201 "uuid": "93db4cea-47aa-455a-a128-165da801217e", 00:07:02.201 "is_configured": true, 00:07:02.201 "data_offset": 0, 00:07:02.201 "data_size": 65536 00:07:02.201 } 00:07:02.201 ] 00:07:02.201 }' 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:02.201 14:06:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:02.459 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.459 [2024-09-30 14:06:07.100553] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:02.718 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.718 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:02.718 "name": "Existed_Raid", 00:07:02.718 "aliases": [ 00:07:02.718 "9d3bae51-f7e9-4f8f-8eb4-732b6ecf20dd" 00:07:02.718 ], 00:07:02.718 "product_name": "Raid Volume", 00:07:02.718 "block_size": 512, 00:07:02.718 "num_blocks": 131072, 00:07:02.718 "uuid": "9d3bae51-f7e9-4f8f-8eb4-732b6ecf20dd", 00:07:02.718 "assigned_rate_limits": { 00:07:02.718 "rw_ios_per_sec": 0, 00:07:02.718 "rw_mbytes_per_sec": 0, 00:07:02.719 "r_mbytes_per_sec": 0, 00:07:02.719 "w_mbytes_per_sec": 0 00:07:02.719 }, 00:07:02.719 "claimed": false, 00:07:02.719 "zoned": false, 00:07:02.719 "supported_io_types": { 00:07:02.719 "read": true, 00:07:02.719 "write": true, 00:07:02.719 "unmap": true, 00:07:02.719 "flush": true, 00:07:02.719 "reset": true, 00:07:02.719 "nvme_admin": false, 00:07:02.719 "nvme_io": false, 00:07:02.719 "nvme_io_md": false, 00:07:02.719 "write_zeroes": true, 00:07:02.719 "zcopy": false, 00:07:02.719 "get_zone_info": false, 00:07:02.719 "zone_management": false, 00:07:02.719 "zone_append": false, 00:07:02.719 "compare": false, 00:07:02.719 "compare_and_write": false, 00:07:02.719 "abort": false, 00:07:02.719 "seek_hole": false, 00:07:02.719 "seek_data": false, 00:07:02.719 "copy": false, 00:07:02.719 "nvme_iov_md": false 00:07:02.719 }, 00:07:02.719 "memory_domains": [ 00:07:02.719 { 00:07:02.719 "dma_device_id": "system", 00:07:02.719 "dma_device_type": 1 00:07:02.719 }, 00:07:02.719 { 00:07:02.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:02.719 "dma_device_type": 2 00:07:02.719 }, 00:07:02.719 { 00:07:02.719 "dma_device_id": "system", 00:07:02.719 "dma_device_type": 1 00:07:02.719 }, 00:07:02.719 { 00:07:02.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:02.719 "dma_device_type": 2 00:07:02.719 } 00:07:02.719 ], 00:07:02.719 "driver_specific": { 00:07:02.719 "raid": { 00:07:02.719 "uuid": "9d3bae51-f7e9-4f8f-8eb4-732b6ecf20dd", 00:07:02.719 "strip_size_kb": 64, 00:07:02.719 "state": "online", 00:07:02.719 "raid_level": "raid0", 00:07:02.719 "superblock": false, 00:07:02.719 "num_base_bdevs": 2, 00:07:02.719 "num_base_bdevs_discovered": 2, 00:07:02.719 "num_base_bdevs_operational": 2, 00:07:02.719 "base_bdevs_list": [ 00:07:02.719 { 00:07:02.719 "name": "BaseBdev1", 00:07:02.719 "uuid": "f13be733-dbe5-4806-9f39-12072cd679ac", 00:07:02.719 "is_configured": true, 00:07:02.719 "data_offset": 0, 00:07:02.719 "data_size": 65536 00:07:02.719 }, 00:07:02.719 { 00:07:02.719 "name": "BaseBdev2", 00:07:02.719 "uuid": "93db4cea-47aa-455a-a128-165da801217e", 00:07:02.719 "is_configured": true, 00:07:02.719 "data_offset": 0, 00:07:02.719 "data_size": 65536 00:07:02.719 } 00:07:02.719 ] 00:07:02.719 } 00:07:02.719 } 00:07:02.719 }' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:02.719 BaseBdev2' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.719 [2024-09-30 14:06:07.276032] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:02.719 [2024-09-30 14:06:07.276059] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:02.719 [2024-09-30 14:06:07.276113] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:02.719 "name": "Existed_Raid", 00:07:02.719 "uuid": "9d3bae51-f7e9-4f8f-8eb4-732b6ecf20dd", 00:07:02.719 "strip_size_kb": 64, 00:07:02.719 "state": "offline", 00:07:02.719 "raid_level": "raid0", 00:07:02.719 "superblock": false, 00:07:02.719 "num_base_bdevs": 2, 00:07:02.719 "num_base_bdevs_discovered": 1, 00:07:02.719 "num_base_bdevs_operational": 1, 00:07:02.719 "base_bdevs_list": [ 00:07:02.719 { 00:07:02.719 "name": null, 00:07:02.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:02.719 "is_configured": false, 00:07:02.719 "data_offset": 0, 00:07:02.719 "data_size": 65536 00:07:02.719 }, 00:07:02.719 { 00:07:02.719 "name": "BaseBdev2", 00:07:02.719 "uuid": "93db4cea-47aa-455a-a128-165da801217e", 00:07:02.719 "is_configured": true, 00:07:02.719 "data_offset": 0, 00:07:02.719 "data_size": 65536 00:07:02.719 } 00:07:02.719 ] 00:07:02.719 }' 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:02.719 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.288 [2024-09-30 14:06:07.734607] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:03.288 [2024-09-30 14:06:07.734668] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73165 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 73165 ']' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 73165 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73165 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:03.288 killing process with pid 73165 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73165' 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 73165 00:07:03.288 [2024-09-30 14:06:07.838990] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:03.288 14:06:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 73165 00:07:03.288 [2024-09-30 14:06:07.839951] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:03.549 00:07:03.549 real 0m3.682s 00:07:03.549 user 0m5.726s 00:07:03.549 sys 0m0.762s 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.549 ************************************ 00:07:03.549 END TEST raid_state_function_test 00:07:03.549 ************************************ 00:07:03.549 14:06:08 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:03.549 14:06:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:03.549 14:06:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:03.549 14:06:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:03.549 ************************************ 00:07:03.549 START TEST raid_state_function_test_sb 00:07:03.549 ************************************ 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 true 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73406 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73406' 00:07:03.549 Process raid pid: 73406 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73406 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 73406 ']' 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:03.549 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:03.549 14:06:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.808 [2024-09-30 14:06:08.239465] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:03.808 [2024-09-30 14:06:08.239583] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:03.808 [2024-09-30 14:06:08.371556] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:03.808 [2024-09-30 14:06:08.401504] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:03.808 [2024-09-30 14:06:08.445714] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:04.067 [2024-09-30 14:06:08.486576] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:04.067 [2024-09-30 14:06:08.486608] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.636 [2024-09-30 14:06:09.067138] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:04.636 [2024-09-30 14:06:09.067195] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:04.636 [2024-09-30 14:06:09.067209] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:04.636 [2024-09-30 14:06:09.067216] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.636 "name": "Existed_Raid", 00:07:04.636 "uuid": "17208aca-2ec5-4df7-a704-cdfd0324ba9b", 00:07:04.636 "strip_size_kb": 64, 00:07:04.636 "state": "configuring", 00:07:04.636 "raid_level": "raid0", 00:07:04.636 "superblock": true, 00:07:04.636 "num_base_bdevs": 2, 00:07:04.636 "num_base_bdevs_discovered": 0, 00:07:04.636 "num_base_bdevs_operational": 2, 00:07:04.636 "base_bdevs_list": [ 00:07:04.636 { 00:07:04.636 "name": "BaseBdev1", 00:07:04.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.636 "is_configured": false, 00:07:04.636 "data_offset": 0, 00:07:04.636 "data_size": 0 00:07:04.636 }, 00:07:04.636 { 00:07:04.636 "name": "BaseBdev2", 00:07:04.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.636 "is_configured": false, 00:07:04.636 "data_offset": 0, 00:07:04.636 "data_size": 0 00:07:04.636 } 00:07:04.636 ] 00:07:04.636 }' 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.636 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 [2024-09-30 14:06:09.450390] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:04.897 [2024-09-30 14:06:09.450426] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 [2024-09-30 14:06:09.458384] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:04.897 [2024-09-30 14:06:09.458421] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:04.897 [2024-09-30 14:06:09.458431] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:04.897 [2024-09-30 14:06:09.458438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 [2024-09-30 14:06:09.475050] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:04.897 BaseBdev1 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 [ 00:07:04.897 { 00:07:04.897 "name": "BaseBdev1", 00:07:04.897 "aliases": [ 00:07:04.897 "88303c10-fd83-452a-8c6d-ed69acd1a5e6" 00:07:04.897 ], 00:07:04.897 "product_name": "Malloc disk", 00:07:04.897 "block_size": 512, 00:07:04.897 "num_blocks": 65536, 00:07:04.897 "uuid": "88303c10-fd83-452a-8c6d-ed69acd1a5e6", 00:07:04.897 "assigned_rate_limits": { 00:07:04.897 "rw_ios_per_sec": 0, 00:07:04.897 "rw_mbytes_per_sec": 0, 00:07:04.897 "r_mbytes_per_sec": 0, 00:07:04.897 "w_mbytes_per_sec": 0 00:07:04.897 }, 00:07:04.897 "claimed": true, 00:07:04.897 "claim_type": "exclusive_write", 00:07:04.897 "zoned": false, 00:07:04.897 "supported_io_types": { 00:07:04.897 "read": true, 00:07:04.897 "write": true, 00:07:04.897 "unmap": true, 00:07:04.897 "flush": true, 00:07:04.897 "reset": true, 00:07:04.897 "nvme_admin": false, 00:07:04.897 "nvme_io": false, 00:07:04.897 "nvme_io_md": false, 00:07:04.897 "write_zeroes": true, 00:07:04.897 "zcopy": true, 00:07:04.897 "get_zone_info": false, 00:07:04.897 "zone_management": false, 00:07:04.897 "zone_append": false, 00:07:04.897 "compare": false, 00:07:04.897 "compare_and_write": false, 00:07:04.897 "abort": true, 00:07:04.897 "seek_hole": false, 00:07:04.897 "seek_data": false, 00:07:04.897 "copy": true, 00:07:04.897 "nvme_iov_md": false 00:07:04.897 }, 00:07:04.897 "memory_domains": [ 00:07:04.897 { 00:07:04.897 "dma_device_id": "system", 00:07:04.897 "dma_device_type": 1 00:07:04.897 }, 00:07:04.897 { 00:07:04.897 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:04.897 "dma_device_type": 2 00:07:04.897 } 00:07:04.897 ], 00:07:04.897 "driver_specific": {} 00:07:04.897 } 00:07:04.897 ] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.897 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.157 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:05.157 "name": "Existed_Raid", 00:07:05.157 "uuid": "a7d85cc0-49a3-40e9-b4cf-0698e6e06fa9", 00:07:05.157 "strip_size_kb": 64, 00:07:05.157 "state": "configuring", 00:07:05.157 "raid_level": "raid0", 00:07:05.157 "superblock": true, 00:07:05.157 "num_base_bdevs": 2, 00:07:05.157 "num_base_bdevs_discovered": 1, 00:07:05.157 "num_base_bdevs_operational": 2, 00:07:05.157 "base_bdevs_list": [ 00:07:05.157 { 00:07:05.157 "name": "BaseBdev1", 00:07:05.157 "uuid": "88303c10-fd83-452a-8c6d-ed69acd1a5e6", 00:07:05.157 "is_configured": true, 00:07:05.158 "data_offset": 2048, 00:07:05.158 "data_size": 63488 00:07:05.158 }, 00:07:05.158 { 00:07:05.158 "name": "BaseBdev2", 00:07:05.158 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:05.158 "is_configured": false, 00:07:05.158 "data_offset": 0, 00:07:05.158 "data_size": 0 00:07:05.158 } 00:07:05.158 ] 00:07:05.158 }' 00:07:05.158 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:05.158 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.417 [2024-09-30 14:06:09.926388] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:05.417 [2024-09-30 14:06:09.926460] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.417 [2024-09-30 14:06:09.938383] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:05.417 [2024-09-30 14:06:09.940214] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:05.417 [2024-09-30 14:06:09.940257] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.417 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:05.418 "name": "Existed_Raid", 00:07:05.418 "uuid": "779bfbbb-c90f-4cb1-95d3-6ec535afaa50", 00:07:05.418 "strip_size_kb": 64, 00:07:05.418 "state": "configuring", 00:07:05.418 "raid_level": "raid0", 00:07:05.418 "superblock": true, 00:07:05.418 "num_base_bdevs": 2, 00:07:05.418 "num_base_bdevs_discovered": 1, 00:07:05.418 "num_base_bdevs_operational": 2, 00:07:05.418 "base_bdevs_list": [ 00:07:05.418 { 00:07:05.418 "name": "BaseBdev1", 00:07:05.418 "uuid": "88303c10-fd83-452a-8c6d-ed69acd1a5e6", 00:07:05.418 "is_configured": true, 00:07:05.418 "data_offset": 2048, 00:07:05.418 "data_size": 63488 00:07:05.418 }, 00:07:05.418 { 00:07:05.418 "name": "BaseBdev2", 00:07:05.418 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:05.418 "is_configured": false, 00:07:05.418 "data_offset": 0, 00:07:05.418 "data_size": 0 00:07:05.418 } 00:07:05.418 ] 00:07:05.418 }' 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:05.418 14:06:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.676 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:05.676 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.676 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.935 [2024-09-30 14:06:10.347784] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:05.935 [2024-09-30 14:06:10.348446] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:05.935 [2024-09-30 14:06:10.348663] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:05.935 BaseBdev2 00:07:05.935 [2024-09-30 14:06:10.349633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.935 [2024-09-30 14:06:10.350132] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:05.935 [2024-09-30 14:06:10.350287] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:05.935 [2024-09-30 14:06:10.350818] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.935 [ 00:07:05.935 { 00:07:05.935 "name": "BaseBdev2", 00:07:05.935 "aliases": [ 00:07:05.935 "e4ce451c-9089-4b50-a965-a5a5220a2914" 00:07:05.935 ], 00:07:05.935 "product_name": "Malloc disk", 00:07:05.935 "block_size": 512, 00:07:05.935 "num_blocks": 65536, 00:07:05.935 "uuid": "e4ce451c-9089-4b50-a965-a5a5220a2914", 00:07:05.935 "assigned_rate_limits": { 00:07:05.935 "rw_ios_per_sec": 0, 00:07:05.935 "rw_mbytes_per_sec": 0, 00:07:05.935 "r_mbytes_per_sec": 0, 00:07:05.935 "w_mbytes_per_sec": 0 00:07:05.935 }, 00:07:05.935 "claimed": true, 00:07:05.935 "claim_type": "exclusive_write", 00:07:05.935 "zoned": false, 00:07:05.935 "supported_io_types": { 00:07:05.935 "read": true, 00:07:05.935 "write": true, 00:07:05.935 "unmap": true, 00:07:05.935 "flush": true, 00:07:05.935 "reset": true, 00:07:05.935 "nvme_admin": false, 00:07:05.935 "nvme_io": false, 00:07:05.935 "nvme_io_md": false, 00:07:05.935 "write_zeroes": true, 00:07:05.935 "zcopy": true, 00:07:05.935 "get_zone_info": false, 00:07:05.935 "zone_management": false, 00:07:05.935 "zone_append": false, 00:07:05.935 "compare": false, 00:07:05.935 "compare_and_write": false, 00:07:05.935 "abort": true, 00:07:05.935 "seek_hole": false, 00:07:05.935 "seek_data": false, 00:07:05.935 "copy": true, 00:07:05.935 "nvme_iov_md": false 00:07:05.935 }, 00:07:05.935 "memory_domains": [ 00:07:05.935 { 00:07:05.935 "dma_device_id": "system", 00:07:05.935 "dma_device_type": 1 00:07:05.935 }, 00:07:05.935 { 00:07:05.935 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.935 "dma_device_type": 2 00:07:05.935 } 00:07:05.935 ], 00:07:05.935 "driver_specific": {} 00:07:05.935 } 00:07:05.935 ] 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:05.935 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:05.936 "name": "Existed_Raid", 00:07:05.936 "uuid": "779bfbbb-c90f-4cb1-95d3-6ec535afaa50", 00:07:05.936 "strip_size_kb": 64, 00:07:05.936 "state": "online", 00:07:05.936 "raid_level": "raid0", 00:07:05.936 "superblock": true, 00:07:05.936 "num_base_bdevs": 2, 00:07:05.936 "num_base_bdevs_discovered": 2, 00:07:05.936 "num_base_bdevs_operational": 2, 00:07:05.936 "base_bdevs_list": [ 00:07:05.936 { 00:07:05.936 "name": "BaseBdev1", 00:07:05.936 "uuid": "88303c10-fd83-452a-8c6d-ed69acd1a5e6", 00:07:05.936 "is_configured": true, 00:07:05.936 "data_offset": 2048, 00:07:05.936 "data_size": 63488 00:07:05.936 }, 00:07:05.936 { 00:07:05.936 "name": "BaseBdev2", 00:07:05.936 "uuid": "e4ce451c-9089-4b50-a965-a5a5220a2914", 00:07:05.936 "is_configured": true, 00:07:05.936 "data_offset": 2048, 00:07:05.936 "data_size": 63488 00:07:05.936 } 00:07:05.936 ] 00:07:05.936 }' 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:05.936 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:06.195 [2024-09-30 14:06:10.791283] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.195 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:06.195 "name": "Existed_Raid", 00:07:06.195 "aliases": [ 00:07:06.196 "779bfbbb-c90f-4cb1-95d3-6ec535afaa50" 00:07:06.196 ], 00:07:06.196 "product_name": "Raid Volume", 00:07:06.196 "block_size": 512, 00:07:06.196 "num_blocks": 126976, 00:07:06.196 "uuid": "779bfbbb-c90f-4cb1-95d3-6ec535afaa50", 00:07:06.196 "assigned_rate_limits": { 00:07:06.196 "rw_ios_per_sec": 0, 00:07:06.196 "rw_mbytes_per_sec": 0, 00:07:06.196 "r_mbytes_per_sec": 0, 00:07:06.196 "w_mbytes_per_sec": 0 00:07:06.196 }, 00:07:06.196 "claimed": false, 00:07:06.196 "zoned": false, 00:07:06.196 "supported_io_types": { 00:07:06.196 "read": true, 00:07:06.196 "write": true, 00:07:06.196 "unmap": true, 00:07:06.196 "flush": true, 00:07:06.196 "reset": true, 00:07:06.196 "nvme_admin": false, 00:07:06.196 "nvme_io": false, 00:07:06.196 "nvme_io_md": false, 00:07:06.196 "write_zeroes": true, 00:07:06.196 "zcopy": false, 00:07:06.196 "get_zone_info": false, 00:07:06.196 "zone_management": false, 00:07:06.196 "zone_append": false, 00:07:06.196 "compare": false, 00:07:06.196 "compare_and_write": false, 00:07:06.196 "abort": false, 00:07:06.196 "seek_hole": false, 00:07:06.196 "seek_data": false, 00:07:06.196 "copy": false, 00:07:06.196 "nvme_iov_md": false 00:07:06.196 }, 00:07:06.196 "memory_domains": [ 00:07:06.196 { 00:07:06.196 "dma_device_id": "system", 00:07:06.196 "dma_device_type": 1 00:07:06.196 }, 00:07:06.196 { 00:07:06.196 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:06.196 "dma_device_type": 2 00:07:06.196 }, 00:07:06.196 { 00:07:06.196 "dma_device_id": "system", 00:07:06.196 "dma_device_type": 1 00:07:06.196 }, 00:07:06.196 { 00:07:06.196 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:06.196 "dma_device_type": 2 00:07:06.196 } 00:07:06.196 ], 00:07:06.196 "driver_specific": { 00:07:06.196 "raid": { 00:07:06.196 "uuid": "779bfbbb-c90f-4cb1-95d3-6ec535afaa50", 00:07:06.196 "strip_size_kb": 64, 00:07:06.196 "state": "online", 00:07:06.196 "raid_level": "raid0", 00:07:06.196 "superblock": true, 00:07:06.196 "num_base_bdevs": 2, 00:07:06.196 "num_base_bdevs_discovered": 2, 00:07:06.196 "num_base_bdevs_operational": 2, 00:07:06.196 "base_bdevs_list": [ 00:07:06.196 { 00:07:06.196 "name": "BaseBdev1", 00:07:06.196 "uuid": "88303c10-fd83-452a-8c6d-ed69acd1a5e6", 00:07:06.196 "is_configured": true, 00:07:06.196 "data_offset": 2048, 00:07:06.196 "data_size": 63488 00:07:06.196 }, 00:07:06.196 { 00:07:06.196 "name": "BaseBdev2", 00:07:06.196 "uuid": "e4ce451c-9089-4b50-a965-a5a5220a2914", 00:07:06.196 "is_configured": true, 00:07:06.196 "data_offset": 2048, 00:07:06.196 "data_size": 63488 00:07:06.196 } 00:07:06.196 ] 00:07:06.196 } 00:07:06.196 } 00:07:06.196 }' 00:07:06.196 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:06.455 BaseBdev2' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.455 14:06:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.455 [2024-09-30 14:06:11.030642] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:06.455 [2024-09-30 14:06:11.030707] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:06.455 [2024-09-30 14:06:11.030780] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.455 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:06.455 "name": "Existed_Raid", 00:07:06.455 "uuid": "779bfbbb-c90f-4cb1-95d3-6ec535afaa50", 00:07:06.455 "strip_size_kb": 64, 00:07:06.455 "state": "offline", 00:07:06.455 "raid_level": "raid0", 00:07:06.455 "superblock": true, 00:07:06.455 "num_base_bdevs": 2, 00:07:06.455 "num_base_bdevs_discovered": 1, 00:07:06.455 "num_base_bdevs_operational": 1, 00:07:06.455 "base_bdevs_list": [ 00:07:06.455 { 00:07:06.455 "name": null, 00:07:06.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:06.455 "is_configured": false, 00:07:06.455 "data_offset": 0, 00:07:06.455 "data_size": 63488 00:07:06.455 }, 00:07:06.455 { 00:07:06.455 "name": "BaseBdev2", 00:07:06.455 "uuid": "e4ce451c-9089-4b50-a965-a5a5220a2914", 00:07:06.455 "is_configured": true, 00:07:06.455 "data_offset": 2048, 00:07:06.455 "data_size": 63488 00:07:06.456 } 00:07:06.456 ] 00:07:06.456 }' 00:07:06.456 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:06.456 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:07.024 [2024-09-30 14:06:11.489015] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:07.024 [2024-09-30 14:06:11.489120] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73406 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 73406 ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 73406 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73406 00:07:07.024 killing process with pid 73406 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73406' 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 73406 00:07:07.024 [2024-09-30 14:06:11.585988] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:07.024 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 73406 00:07:07.024 [2024-09-30 14:06:11.586971] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:07.283 ************************************ 00:07:07.283 END TEST raid_state_function_test_sb 00:07:07.283 ************************************ 00:07:07.283 14:06:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:07.283 00:07:07.283 real 0m3.684s 00:07:07.283 user 0m5.762s 00:07:07.283 sys 0m0.719s 00:07:07.283 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:07.283 14:06:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:07.283 14:06:11 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:07.283 14:06:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:07.283 14:06:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:07.283 14:06:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:07.283 ************************************ 00:07:07.283 START TEST raid_superblock_test 00:07:07.283 ************************************ 00:07:07.283 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 2 00:07:07.283 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:07.283 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73643 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73643 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 73643 ']' 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:07.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:07.284 14:06:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.543 [2024-09-30 14:06:11.989946] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:07.543 [2024-09-30 14:06:11.990172] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73643 ] 00:07:07.543 [2024-09-30 14:06:12.121066] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:07.543 [2024-09-30 14:06:12.150748] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.543 [2024-09-30 14:06:12.197874] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.802 [2024-09-30 14:06:12.240668] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:07.802 [2024-09-30 14:06:12.240776] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.370 malloc1 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.370 [2024-09-30 14:06:12.835005] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:08.370 [2024-09-30 14:06:12.835110] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.370 [2024-09-30 14:06:12.835155] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:08.370 [2024-09-30 14:06:12.835221] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.370 [2024-09-30 14:06:12.837231] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.370 [2024-09-30 14:06:12.837301] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:08.370 pt1 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.370 malloc2 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.370 [2024-09-30 14:06:12.883425] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:08.370 [2024-09-30 14:06:12.883578] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.370 [2024-09-30 14:06:12.883628] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:08.370 [2024-09-30 14:06:12.883654] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.370 [2024-09-30 14:06:12.887901] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.370 [2024-09-30 14:06:12.887956] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:08.370 pt2 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.370 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.371 [2024-09-30 14:06:12.896206] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:08.371 [2024-09-30 14:06:12.898586] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:08.371 [2024-09-30 14:06:12.898759] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:07:08.371 [2024-09-30 14:06:12.898781] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:08.371 [2024-09-30 14:06:12.899082] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:08.371 [2024-09-30 14:06:12.899247] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:07:08.371 [2024-09-30 14:06:12.899263] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:07:08.371 [2024-09-30 14:06:12.899424] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:08.371 "name": "raid_bdev1", 00:07:08.371 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:08.371 "strip_size_kb": 64, 00:07:08.371 "state": "online", 00:07:08.371 "raid_level": "raid0", 00:07:08.371 "superblock": true, 00:07:08.371 "num_base_bdevs": 2, 00:07:08.371 "num_base_bdevs_discovered": 2, 00:07:08.371 "num_base_bdevs_operational": 2, 00:07:08.371 "base_bdevs_list": [ 00:07:08.371 { 00:07:08.371 "name": "pt1", 00:07:08.371 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.371 "is_configured": true, 00:07:08.371 "data_offset": 2048, 00:07:08.371 "data_size": 63488 00:07:08.371 }, 00:07:08.371 { 00:07:08.371 "name": "pt2", 00:07:08.371 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.371 "is_configured": true, 00:07:08.371 "data_offset": 2048, 00:07:08.371 "data_size": 63488 00:07:08.371 } 00:07:08.371 ] 00:07:08.371 }' 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:08.371 14:06:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.940 [2024-09-30 14:06:13.311689] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:08.940 "name": "raid_bdev1", 00:07:08.940 "aliases": [ 00:07:08.940 "bced43a1-beb7-4521-ad7e-652c421a0919" 00:07:08.940 ], 00:07:08.940 "product_name": "Raid Volume", 00:07:08.940 "block_size": 512, 00:07:08.940 "num_blocks": 126976, 00:07:08.940 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:08.940 "assigned_rate_limits": { 00:07:08.940 "rw_ios_per_sec": 0, 00:07:08.940 "rw_mbytes_per_sec": 0, 00:07:08.940 "r_mbytes_per_sec": 0, 00:07:08.940 "w_mbytes_per_sec": 0 00:07:08.940 }, 00:07:08.940 "claimed": false, 00:07:08.940 "zoned": false, 00:07:08.940 "supported_io_types": { 00:07:08.940 "read": true, 00:07:08.940 "write": true, 00:07:08.940 "unmap": true, 00:07:08.940 "flush": true, 00:07:08.940 "reset": true, 00:07:08.940 "nvme_admin": false, 00:07:08.940 "nvme_io": false, 00:07:08.940 "nvme_io_md": false, 00:07:08.940 "write_zeroes": true, 00:07:08.940 "zcopy": false, 00:07:08.940 "get_zone_info": false, 00:07:08.940 "zone_management": false, 00:07:08.940 "zone_append": false, 00:07:08.940 "compare": false, 00:07:08.940 "compare_and_write": false, 00:07:08.940 "abort": false, 00:07:08.940 "seek_hole": false, 00:07:08.940 "seek_data": false, 00:07:08.940 "copy": false, 00:07:08.940 "nvme_iov_md": false 00:07:08.940 }, 00:07:08.940 "memory_domains": [ 00:07:08.940 { 00:07:08.940 "dma_device_id": "system", 00:07:08.940 "dma_device_type": 1 00:07:08.940 }, 00:07:08.940 { 00:07:08.940 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.940 "dma_device_type": 2 00:07:08.940 }, 00:07:08.940 { 00:07:08.940 "dma_device_id": "system", 00:07:08.940 "dma_device_type": 1 00:07:08.940 }, 00:07:08.940 { 00:07:08.940 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.940 "dma_device_type": 2 00:07:08.940 } 00:07:08.940 ], 00:07:08.940 "driver_specific": { 00:07:08.940 "raid": { 00:07:08.940 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:08.940 "strip_size_kb": 64, 00:07:08.940 "state": "online", 00:07:08.940 "raid_level": "raid0", 00:07:08.940 "superblock": true, 00:07:08.940 "num_base_bdevs": 2, 00:07:08.940 "num_base_bdevs_discovered": 2, 00:07:08.940 "num_base_bdevs_operational": 2, 00:07:08.940 "base_bdevs_list": [ 00:07:08.940 { 00:07:08.940 "name": "pt1", 00:07:08.940 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.940 "is_configured": true, 00:07:08.940 "data_offset": 2048, 00:07:08.940 "data_size": 63488 00:07:08.940 }, 00:07:08.940 { 00:07:08.940 "name": "pt2", 00:07:08.940 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.940 "is_configured": true, 00:07:08.940 "data_offset": 2048, 00:07:08.940 "data_size": 63488 00:07:08.940 } 00:07:08.940 ] 00:07:08.940 } 00:07:08.940 } 00:07:08.940 }' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:08.940 pt2' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.940 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.941 [2024-09-30 14:06:13.543192] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bced43a1-beb7-4521-ad7e-652c421a0919 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z bced43a1-beb7-4521-ad7e-652c421a0919 ']' 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.941 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.941 [2024-09-30 14:06:13.590866] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:08.941 [2024-09-30 14:06:13.590924] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:08.941 [2024-09-30 14:06:13.591013] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:08.941 [2024-09-30 14:06:13.591084] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:08.941 [2024-09-30 14:06:13.591121] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 [2024-09-30 14:06:13.726659] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:09.200 [2024-09-30 14:06:13.728424] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:09.200 [2024-09-30 14:06:13.728480] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:09.200 [2024-09-30 14:06:13.728560] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:09.200 [2024-09-30 14:06:13.728576] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:09.200 [2024-09-30 14:06:13.728586] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:07:09.200 request: 00:07:09.200 { 00:07:09.200 "name": "raid_bdev1", 00:07:09.200 "raid_level": "raid0", 00:07:09.200 "base_bdevs": [ 00:07:09.200 "malloc1", 00:07:09.200 "malloc2" 00:07:09.200 ], 00:07:09.200 "strip_size_kb": 64, 00:07:09.200 "superblock": false, 00:07:09.200 "method": "bdev_raid_create", 00:07:09.200 "req_id": 1 00:07:09.200 } 00:07:09.200 Got JSON-RPC error response 00:07:09.200 response: 00:07:09.200 { 00:07:09.200 "code": -17, 00:07:09.200 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:09.200 } 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.200 [2024-09-30 14:06:13.794571] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:09.200 [2024-09-30 14:06:13.794654] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:09.200 [2024-09-30 14:06:13.794684] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:09.200 [2024-09-30 14:06:13.794713] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:09.200 [2024-09-30 14:06:13.796727] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:09.200 [2024-09-30 14:06:13.796796] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:09.200 [2024-09-30 14:06:13.796875] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:09.200 [2024-09-30 14:06:13.796944] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:09.200 pt1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:09.200 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:09.201 "name": "raid_bdev1", 00:07:09.201 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:09.201 "strip_size_kb": 64, 00:07:09.201 "state": "configuring", 00:07:09.201 "raid_level": "raid0", 00:07:09.201 "superblock": true, 00:07:09.201 "num_base_bdevs": 2, 00:07:09.201 "num_base_bdevs_discovered": 1, 00:07:09.201 "num_base_bdevs_operational": 2, 00:07:09.201 "base_bdevs_list": [ 00:07:09.201 { 00:07:09.201 "name": "pt1", 00:07:09.201 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:09.201 "is_configured": true, 00:07:09.201 "data_offset": 2048, 00:07:09.201 "data_size": 63488 00:07:09.201 }, 00:07:09.201 { 00:07:09.201 "name": null, 00:07:09.201 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:09.201 "is_configured": false, 00:07:09.201 "data_offset": 2048, 00:07:09.201 "data_size": 63488 00:07:09.201 } 00:07:09.201 ] 00:07:09.201 }' 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:09.201 14:06:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.770 [2024-09-30 14:06:14.197914] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:09.770 [2024-09-30 14:06:14.197974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:09.770 [2024-09-30 14:06:14.197995] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:09.770 [2024-09-30 14:06:14.198005] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:09.770 [2024-09-30 14:06:14.198381] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:09.770 [2024-09-30 14:06:14.198400] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:09.770 [2024-09-30 14:06:14.198476] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:09.770 [2024-09-30 14:06:14.198533] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:09.770 [2024-09-30 14:06:14.198623] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:09.770 [2024-09-30 14:06:14.198639] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:09.770 [2024-09-30 14:06:14.198867] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:09.770 [2024-09-30 14:06:14.198985] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:09.770 [2024-09-30 14:06:14.198995] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:09.770 [2024-09-30 14:06:14.199093] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:09.770 pt2 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:09.770 "name": "raid_bdev1", 00:07:09.770 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:09.770 "strip_size_kb": 64, 00:07:09.770 "state": "online", 00:07:09.770 "raid_level": "raid0", 00:07:09.770 "superblock": true, 00:07:09.770 "num_base_bdevs": 2, 00:07:09.770 "num_base_bdevs_discovered": 2, 00:07:09.770 "num_base_bdevs_operational": 2, 00:07:09.770 "base_bdevs_list": [ 00:07:09.770 { 00:07:09.770 "name": "pt1", 00:07:09.770 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:09.770 "is_configured": true, 00:07:09.770 "data_offset": 2048, 00:07:09.770 "data_size": 63488 00:07:09.770 }, 00:07:09.770 { 00:07:09.770 "name": "pt2", 00:07:09.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:09.770 "is_configured": true, 00:07:09.770 "data_offset": 2048, 00:07:09.770 "data_size": 63488 00:07:09.770 } 00:07:09.770 ] 00:07:09.770 }' 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:09.770 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.029 [2024-09-30 14:06:14.617438] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.029 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:10.029 "name": "raid_bdev1", 00:07:10.029 "aliases": [ 00:07:10.029 "bced43a1-beb7-4521-ad7e-652c421a0919" 00:07:10.030 ], 00:07:10.030 "product_name": "Raid Volume", 00:07:10.030 "block_size": 512, 00:07:10.030 "num_blocks": 126976, 00:07:10.030 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:10.030 "assigned_rate_limits": { 00:07:10.030 "rw_ios_per_sec": 0, 00:07:10.030 "rw_mbytes_per_sec": 0, 00:07:10.030 "r_mbytes_per_sec": 0, 00:07:10.030 "w_mbytes_per_sec": 0 00:07:10.030 }, 00:07:10.030 "claimed": false, 00:07:10.030 "zoned": false, 00:07:10.030 "supported_io_types": { 00:07:10.030 "read": true, 00:07:10.030 "write": true, 00:07:10.030 "unmap": true, 00:07:10.030 "flush": true, 00:07:10.030 "reset": true, 00:07:10.030 "nvme_admin": false, 00:07:10.030 "nvme_io": false, 00:07:10.030 "nvme_io_md": false, 00:07:10.030 "write_zeroes": true, 00:07:10.030 "zcopy": false, 00:07:10.030 "get_zone_info": false, 00:07:10.030 "zone_management": false, 00:07:10.030 "zone_append": false, 00:07:10.030 "compare": false, 00:07:10.030 "compare_and_write": false, 00:07:10.030 "abort": false, 00:07:10.030 "seek_hole": false, 00:07:10.030 "seek_data": false, 00:07:10.030 "copy": false, 00:07:10.030 "nvme_iov_md": false 00:07:10.030 }, 00:07:10.030 "memory_domains": [ 00:07:10.030 { 00:07:10.030 "dma_device_id": "system", 00:07:10.030 "dma_device_type": 1 00:07:10.030 }, 00:07:10.030 { 00:07:10.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:10.030 "dma_device_type": 2 00:07:10.030 }, 00:07:10.030 { 00:07:10.030 "dma_device_id": "system", 00:07:10.030 "dma_device_type": 1 00:07:10.030 }, 00:07:10.030 { 00:07:10.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:10.030 "dma_device_type": 2 00:07:10.030 } 00:07:10.030 ], 00:07:10.030 "driver_specific": { 00:07:10.030 "raid": { 00:07:10.030 "uuid": "bced43a1-beb7-4521-ad7e-652c421a0919", 00:07:10.030 "strip_size_kb": 64, 00:07:10.030 "state": "online", 00:07:10.030 "raid_level": "raid0", 00:07:10.030 "superblock": true, 00:07:10.030 "num_base_bdevs": 2, 00:07:10.030 "num_base_bdevs_discovered": 2, 00:07:10.030 "num_base_bdevs_operational": 2, 00:07:10.030 "base_bdevs_list": [ 00:07:10.030 { 00:07:10.030 "name": "pt1", 00:07:10.030 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:10.030 "is_configured": true, 00:07:10.030 "data_offset": 2048, 00:07:10.030 "data_size": 63488 00:07:10.030 }, 00:07:10.030 { 00:07:10.030 "name": "pt2", 00:07:10.030 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:10.030 "is_configured": true, 00:07:10.030 "data_offset": 2048, 00:07:10.030 "data_size": 63488 00:07:10.030 } 00:07:10.030 ] 00:07:10.030 } 00:07:10.030 } 00:07:10.030 }' 00:07:10.030 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:10.030 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:10.030 pt2' 00:07:10.030 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.290 [2024-09-30 14:06:14.805092] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' bced43a1-beb7-4521-ad7e-652c421a0919 '!=' bced43a1-beb7-4521-ad7e-652c421a0919 ']' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73643 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 73643 ']' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 73643 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73643 00:07:10.290 killing process with pid 73643 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73643' 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 73643 00:07:10.290 [2024-09-30 14:06:14.892667] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:10.290 [2024-09-30 14:06:14.892747] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:10.290 [2024-09-30 14:06:14.892790] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:10.290 [2024-09-30 14:06:14.892800] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:10.290 14:06:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 73643 00:07:10.290 [2024-09-30 14:06:14.914513] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:10.550 ************************************ 00:07:10.550 END TEST raid_superblock_test 00:07:10.550 ************************************ 00:07:10.550 14:06:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:10.550 00:07:10.550 real 0m3.253s 00:07:10.550 user 0m4.962s 00:07:10.550 sys 0m0.694s 00:07:10.550 14:06:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:10.550 14:06:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.810 14:06:15 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:10.810 14:06:15 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:10.810 14:06:15 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:10.810 14:06:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:10.810 ************************************ 00:07:10.810 START TEST raid_read_error_test 00:07:10.810 ************************************ 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 read 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2qL9Ok5KqI 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73838 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73838 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 73838 ']' 00:07:10.810 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:10.810 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:10.811 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:10.811 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:10.811 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:10.811 14:06:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.811 [2024-09-30 14:06:15.330108] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:10.811 [2024-09-30 14:06:15.330230] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73838 ] 00:07:10.811 [2024-09-30 14:06:15.461042] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:11.070 [2024-09-30 14:06:15.486218] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:11.070 [2024-09-30 14:06:15.529998] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.070 [2024-09-30 14:06:15.571397] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:11.070 [2024-09-30 14:06:15.571431] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.639 BaseBdev1_malloc 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.639 true 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.639 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.639 [2024-09-30 14:06:16.180959] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:11.640 [2024-09-30 14:06:16.181014] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:11.640 [2024-09-30 14:06:16.181059] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:11.640 [2024-09-30 14:06:16.181074] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:11.640 [2024-09-30 14:06:16.183104] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:11.640 [2024-09-30 14:06:16.183140] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:11.640 BaseBdev1 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.640 BaseBdev2_malloc 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.640 true 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.640 [2024-09-30 14:06:16.238979] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:11.640 [2024-09-30 14:06:16.239054] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:11.640 [2024-09-30 14:06:16.239079] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:11.640 [2024-09-30 14:06:16.239096] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:11.640 [2024-09-30 14:06:16.242315] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:11.640 [2024-09-30 14:06:16.242376] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:11.640 BaseBdev2 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.640 [2024-09-30 14:06:16.251063] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:11.640 [2024-09-30 14:06:16.253092] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:11.640 [2024-09-30 14:06:16.253250] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:11.640 [2024-09-30 14:06:16.253265] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:11.640 [2024-09-30 14:06:16.253507] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:11.640 [2024-09-30 14:06:16.253653] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:11.640 [2024-09-30 14:06:16.253663] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:11.640 [2024-09-30 14:06:16.253795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.640 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.900 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:11.900 "name": "raid_bdev1", 00:07:11.900 "uuid": "287a7c5a-244e-408c-8823-029f7b6959f8", 00:07:11.900 "strip_size_kb": 64, 00:07:11.900 "state": "online", 00:07:11.900 "raid_level": "raid0", 00:07:11.900 "superblock": true, 00:07:11.900 "num_base_bdevs": 2, 00:07:11.900 "num_base_bdevs_discovered": 2, 00:07:11.900 "num_base_bdevs_operational": 2, 00:07:11.900 "base_bdevs_list": [ 00:07:11.900 { 00:07:11.900 "name": "BaseBdev1", 00:07:11.900 "uuid": "741730dd-6560-5a8e-b869-668ebe7b005e", 00:07:11.900 "is_configured": true, 00:07:11.900 "data_offset": 2048, 00:07:11.900 "data_size": 63488 00:07:11.900 }, 00:07:11.900 { 00:07:11.900 "name": "BaseBdev2", 00:07:11.900 "uuid": "09e35b90-c9ad-590e-9be3-9fe0d2a63dde", 00:07:11.900 "is_configured": true, 00:07:11.900 "data_offset": 2048, 00:07:11.900 "data_size": 63488 00:07:11.900 } 00:07:11.900 ] 00:07:11.900 }' 00:07:11.900 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:11.900 14:06:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.160 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:12.160 14:06:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:12.160 [2024-09-30 14:06:16.798516] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.099 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.364 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:13.364 "name": "raid_bdev1", 00:07:13.364 "uuid": "287a7c5a-244e-408c-8823-029f7b6959f8", 00:07:13.364 "strip_size_kb": 64, 00:07:13.364 "state": "online", 00:07:13.364 "raid_level": "raid0", 00:07:13.364 "superblock": true, 00:07:13.364 "num_base_bdevs": 2, 00:07:13.364 "num_base_bdevs_discovered": 2, 00:07:13.364 "num_base_bdevs_operational": 2, 00:07:13.364 "base_bdevs_list": [ 00:07:13.364 { 00:07:13.364 "name": "BaseBdev1", 00:07:13.364 "uuid": "741730dd-6560-5a8e-b869-668ebe7b005e", 00:07:13.364 "is_configured": true, 00:07:13.364 "data_offset": 2048, 00:07:13.364 "data_size": 63488 00:07:13.364 }, 00:07:13.364 { 00:07:13.364 "name": "BaseBdev2", 00:07:13.364 "uuid": "09e35b90-c9ad-590e-9be3-9fe0d2a63dde", 00:07:13.364 "is_configured": true, 00:07:13.364 "data_offset": 2048, 00:07:13.364 "data_size": 63488 00:07:13.364 } 00:07:13.364 ] 00:07:13.364 }' 00:07:13.364 14:06:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:13.364 14:06:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.628 [2024-09-30 14:06:18.146099] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:13.628 [2024-09-30 14:06:18.146201] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:13.628 [2024-09-30 14:06:18.148717] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:13.628 [2024-09-30 14:06:18.148798] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:13.628 [2024-09-30 14:06:18.148847] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:13.628 [2024-09-30 14:06:18.148887] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:13.628 { 00:07:13.628 "results": [ 00:07:13.628 { 00:07:13.628 "job": "raid_bdev1", 00:07:13.628 "core_mask": "0x1", 00:07:13.628 "workload": "randrw", 00:07:13.628 "percentage": 50, 00:07:13.628 "status": "finished", 00:07:13.628 "queue_depth": 1, 00:07:13.628 "io_size": 131072, 00:07:13.628 "runtime": 1.348526, 00:07:13.628 "iops": 17866.17388170491, 00:07:13.628 "mibps": 2233.271735213114, 00:07:13.628 "io_failed": 1, 00:07:13.628 "io_timeout": 0, 00:07:13.628 "avg_latency_us": 77.41715080273296, 00:07:13.628 "min_latency_us": 24.482096069868994, 00:07:13.628 "max_latency_us": 1438.071615720524 00:07:13.628 } 00:07:13.628 ], 00:07:13.628 "core_count": 1 00:07:13.628 } 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73838 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 73838 ']' 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 73838 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73838 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:13.628 killing process with pid 73838 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73838' 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 73838 00:07:13.628 [2024-09-30 14:06:18.182383] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:13.628 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 73838 00:07:13.628 [2024-09-30 14:06:18.196998] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2qL9Ok5KqI 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:13.895 ************************************ 00:07:13.895 END TEST raid_read_error_test 00:07:13.895 ************************************ 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:13.895 00:07:13.895 real 0m3.205s 00:07:13.895 user 0m4.054s 00:07:13.895 sys 0m0.492s 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:13.895 14:06:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.895 14:06:18 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:13.895 14:06:18 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:13.895 14:06:18 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:13.895 14:06:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:13.895 ************************************ 00:07:13.895 START TEST raid_write_error_test 00:07:13.895 ************************************ 00:07:13.895 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 write 00:07:13.895 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:13.895 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:13.895 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:13.895 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.oDqo4Xrev4 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73967 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73967 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 73967 ']' 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:13.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:13.896 14:06:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.155 [2024-09-30 14:06:18.620887] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:14.155 [2024-09-30 14:06:18.621015] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73967 ] 00:07:14.155 [2024-09-30 14:06:18.757729] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:14.155 [2024-09-30 14:06:18.786856] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:14.415 [2024-09-30 14:06:18.831690] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:14.415 [2024-09-30 14:06:18.872678] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:14.415 [2024-09-30 14:06:18.872716] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 BaseBdev1_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 true 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 [2024-09-30 14:06:19.454097] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:14.986 [2024-09-30 14:06:19.454154] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:14.986 [2024-09-30 14:06:19.454170] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:14.986 [2024-09-30 14:06:19.454189] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:14.986 [2024-09-30 14:06:19.456232] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:14.986 [2024-09-30 14:06:19.456268] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:14.986 BaseBdev1 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 BaseBdev2_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 true 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.986 [2024-09-30 14:06:19.516361] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:14.986 [2024-09-30 14:06:19.516439] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:14.986 [2024-09-30 14:06:19.516466] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:14.986 [2024-09-30 14:06:19.516503] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:14.986 [2024-09-30 14:06:19.519848] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:14.986 [2024-09-30 14:06:19.519967] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:14.986 BaseBdev2 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.986 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.987 [2024-09-30 14:06:19.528519] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:14.987 [2024-09-30 14:06:19.530570] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:14.987 [2024-09-30 14:06:19.530773] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:14.987 [2024-09-30 14:06:19.530791] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:14.987 [2024-09-30 14:06:19.531049] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:14.987 [2024-09-30 14:06:19.531194] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:14.987 [2024-09-30 14:06:19.531204] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:14.987 [2024-09-30 14:06:19.531332] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:14.987 "name": "raid_bdev1", 00:07:14.987 "uuid": "c624c8ef-a9bf-48ca-9715-e4689c38fb92", 00:07:14.987 "strip_size_kb": 64, 00:07:14.987 "state": "online", 00:07:14.987 "raid_level": "raid0", 00:07:14.987 "superblock": true, 00:07:14.987 "num_base_bdevs": 2, 00:07:14.987 "num_base_bdevs_discovered": 2, 00:07:14.987 "num_base_bdevs_operational": 2, 00:07:14.987 "base_bdevs_list": [ 00:07:14.987 { 00:07:14.987 "name": "BaseBdev1", 00:07:14.987 "uuid": "dafc7559-b13e-50b0-bab0-05fc1ca162bc", 00:07:14.987 "is_configured": true, 00:07:14.987 "data_offset": 2048, 00:07:14.987 "data_size": 63488 00:07:14.987 }, 00:07:14.987 { 00:07:14.987 "name": "BaseBdev2", 00:07:14.987 "uuid": "d0a0465d-999d-5c7d-b5c6-e194d7b874cf", 00:07:14.987 "is_configured": true, 00:07:14.987 "data_offset": 2048, 00:07:14.987 "data_size": 63488 00:07:14.987 } 00:07:14.987 ] 00:07:14.987 }' 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:14.987 14:06:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.555 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:15.555 14:06:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:15.555 [2024-09-30 14:06:20.055957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:16.494 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:16.495 14:06:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.495 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:16.495 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:16.495 "name": "raid_bdev1", 00:07:16.495 "uuid": "c624c8ef-a9bf-48ca-9715-e4689c38fb92", 00:07:16.495 "strip_size_kb": 64, 00:07:16.495 "state": "online", 00:07:16.495 "raid_level": "raid0", 00:07:16.495 "superblock": true, 00:07:16.495 "num_base_bdevs": 2, 00:07:16.495 "num_base_bdevs_discovered": 2, 00:07:16.495 "num_base_bdevs_operational": 2, 00:07:16.495 "base_bdevs_list": [ 00:07:16.495 { 00:07:16.495 "name": "BaseBdev1", 00:07:16.495 "uuid": "dafc7559-b13e-50b0-bab0-05fc1ca162bc", 00:07:16.495 "is_configured": true, 00:07:16.495 "data_offset": 2048, 00:07:16.495 "data_size": 63488 00:07:16.495 }, 00:07:16.495 { 00:07:16.495 "name": "BaseBdev2", 00:07:16.495 "uuid": "d0a0465d-999d-5c7d-b5c6-e194d7b874cf", 00:07:16.495 "is_configured": true, 00:07:16.495 "data_offset": 2048, 00:07:16.495 "data_size": 63488 00:07:16.495 } 00:07:16.495 ] 00:07:16.495 }' 00:07:16.495 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:16.495 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-09-30 14:06:21.439442] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:17.064 [2024-09-30 14:06:21.439558] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:17.064 [2024-09-30 14:06:21.442035] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:17.064 [2024-09-30 14:06:21.442110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:17.064 [2024-09-30 14:06:21.442156] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:17.064 [2024-09-30 14:06:21.442196] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:17.064 { 00:07:17.064 "results": [ 00:07:17.064 { 00:07:17.064 "job": "raid_bdev1", 00:07:17.064 "core_mask": "0x1", 00:07:17.064 "workload": "randrw", 00:07:17.064 "percentage": 50, 00:07:17.064 "status": "finished", 00:07:17.064 "queue_depth": 1, 00:07:17.064 "io_size": 131072, 00:07:17.064 "runtime": 1.384523, 00:07:17.064 "iops": 18077.706184729326, 00:07:17.064 "mibps": 2259.713273091166, 00:07:17.064 "io_failed": 1, 00:07:17.064 "io_timeout": 0, 00:07:17.064 "avg_latency_us": 76.53365257760557, 00:07:17.064 "min_latency_us": 24.258515283842794, 00:07:17.064 "max_latency_us": 1359.3711790393013 00:07:17.064 } 00:07:17.064 ], 00:07:17.064 "core_count": 1 00:07:17.064 } 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73967 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 73967 ']' 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 73967 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73967 00:07:17.064 killing process with pid 73967 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73967' 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 73967 00:07:17.064 [2024-09-30 14:06:21.491049] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:17.064 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 73967 00:07:17.064 [2024-09-30 14:06:21.506404] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.oDqo4Xrev4 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:17.323 00:07:17.323 real 0m3.237s 00:07:17.323 user 0m4.081s 00:07:17.323 sys 0m0.527s 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:17.323 14:06:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.323 ************************************ 00:07:17.323 END TEST raid_write_error_test 00:07:17.323 ************************************ 00:07:17.323 14:06:21 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:17.323 14:06:21 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:17.323 14:06:21 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:17.323 14:06:21 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:17.323 14:06:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:17.323 ************************************ 00:07:17.323 START TEST raid_state_function_test 00:07:17.323 ************************************ 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 false 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74094 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:17.323 Process raid pid: 74094 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74094' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74094 00:07:17.323 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 74094 ']' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:17.323 14:06:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.323 [2024-09-30 14:06:21.914361] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:17.323 [2024-09-30 14:06:21.914584] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:17.582 [2024-09-30 14:06:22.046060] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:17.582 [2024-09-30 14:06:22.074245] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:17.582 [2024-09-30 14:06:22.118754] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.582 [2024-09-30 14:06:22.159288] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.582 [2024-09-30 14:06:22.159393] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.151 [2024-09-30 14:06:22.743886] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:18.151 [2024-09-30 14:06:22.743943] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:18.151 [2024-09-30 14:06:22.743963] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:18.151 [2024-09-30 14:06:22.743971] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.151 "name": "Existed_Raid", 00:07:18.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.151 "strip_size_kb": 64, 00:07:18.151 "state": "configuring", 00:07:18.151 "raid_level": "concat", 00:07:18.151 "superblock": false, 00:07:18.151 "num_base_bdevs": 2, 00:07:18.151 "num_base_bdevs_discovered": 0, 00:07:18.151 "num_base_bdevs_operational": 2, 00:07:18.151 "base_bdevs_list": [ 00:07:18.151 { 00:07:18.151 "name": "BaseBdev1", 00:07:18.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.151 "is_configured": false, 00:07:18.151 "data_offset": 0, 00:07:18.151 "data_size": 0 00:07:18.151 }, 00:07:18.151 { 00:07:18.151 "name": "BaseBdev2", 00:07:18.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.151 "is_configured": false, 00:07:18.151 "data_offset": 0, 00:07:18.151 "data_size": 0 00:07:18.151 } 00:07:18.151 ] 00:07:18.151 }' 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.151 14:06:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 [2024-09-30 14:06:23.167089] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:18.719 [2024-09-30 14:06:23.167174] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 [2024-09-30 14:06:23.175072] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:18.719 [2024-09-30 14:06:23.175144] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:18.719 [2024-09-30 14:06:23.175173] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:18.719 [2024-09-30 14:06:23.175196] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 [2024-09-30 14:06:23.191793] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:18.719 BaseBdev1 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.719 [ 00:07:18.719 { 00:07:18.719 "name": "BaseBdev1", 00:07:18.719 "aliases": [ 00:07:18.719 "07e905ff-b20b-4941-93e7-a6b196c00b40" 00:07:18.719 ], 00:07:18.719 "product_name": "Malloc disk", 00:07:18.719 "block_size": 512, 00:07:18.719 "num_blocks": 65536, 00:07:18.719 "uuid": "07e905ff-b20b-4941-93e7-a6b196c00b40", 00:07:18.719 "assigned_rate_limits": { 00:07:18.719 "rw_ios_per_sec": 0, 00:07:18.719 "rw_mbytes_per_sec": 0, 00:07:18.719 "r_mbytes_per_sec": 0, 00:07:18.719 "w_mbytes_per_sec": 0 00:07:18.719 }, 00:07:18.719 "claimed": true, 00:07:18.719 "claim_type": "exclusive_write", 00:07:18.719 "zoned": false, 00:07:18.719 "supported_io_types": { 00:07:18.719 "read": true, 00:07:18.719 "write": true, 00:07:18.719 "unmap": true, 00:07:18.719 "flush": true, 00:07:18.719 "reset": true, 00:07:18.719 "nvme_admin": false, 00:07:18.719 "nvme_io": false, 00:07:18.719 "nvme_io_md": false, 00:07:18.719 "write_zeroes": true, 00:07:18.719 "zcopy": true, 00:07:18.719 "get_zone_info": false, 00:07:18.719 "zone_management": false, 00:07:18.719 "zone_append": false, 00:07:18.719 "compare": false, 00:07:18.719 "compare_and_write": false, 00:07:18.719 "abort": true, 00:07:18.719 "seek_hole": false, 00:07:18.719 "seek_data": false, 00:07:18.719 "copy": true, 00:07:18.719 "nvme_iov_md": false 00:07:18.719 }, 00:07:18.719 "memory_domains": [ 00:07:18.719 { 00:07:18.719 "dma_device_id": "system", 00:07:18.719 "dma_device_type": 1 00:07:18.719 }, 00:07:18.719 { 00:07:18.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.719 "dma_device_type": 2 00:07:18.719 } 00:07:18.719 ], 00:07:18.719 "driver_specific": {} 00:07:18.719 } 00:07:18.719 ] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.719 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.720 "name": "Existed_Raid", 00:07:18.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.720 "strip_size_kb": 64, 00:07:18.720 "state": "configuring", 00:07:18.720 "raid_level": "concat", 00:07:18.720 "superblock": false, 00:07:18.720 "num_base_bdevs": 2, 00:07:18.720 "num_base_bdevs_discovered": 1, 00:07:18.720 "num_base_bdevs_operational": 2, 00:07:18.720 "base_bdevs_list": [ 00:07:18.720 { 00:07:18.720 "name": "BaseBdev1", 00:07:18.720 "uuid": "07e905ff-b20b-4941-93e7-a6b196c00b40", 00:07:18.720 "is_configured": true, 00:07:18.720 "data_offset": 0, 00:07:18.720 "data_size": 65536 00:07:18.720 }, 00:07:18.720 { 00:07:18.720 "name": "BaseBdev2", 00:07:18.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.720 "is_configured": false, 00:07:18.720 "data_offset": 0, 00:07:18.720 "data_size": 0 00:07:18.720 } 00:07:18.720 ] 00:07:18.720 }' 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.720 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.979 [2024-09-30 14:06:23.619093] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:18.979 [2024-09-30 14:06:23.619179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.979 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.979 [2024-09-30 14:06:23.631100] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:18.979 [2024-09-30 14:06:23.632903] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:18.979 [2024-09-30 14:06:23.632939] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.239 "name": "Existed_Raid", 00:07:19.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.239 "strip_size_kb": 64, 00:07:19.239 "state": "configuring", 00:07:19.239 "raid_level": "concat", 00:07:19.239 "superblock": false, 00:07:19.239 "num_base_bdevs": 2, 00:07:19.239 "num_base_bdevs_discovered": 1, 00:07:19.239 "num_base_bdevs_operational": 2, 00:07:19.239 "base_bdevs_list": [ 00:07:19.239 { 00:07:19.239 "name": "BaseBdev1", 00:07:19.239 "uuid": "07e905ff-b20b-4941-93e7-a6b196c00b40", 00:07:19.239 "is_configured": true, 00:07:19.239 "data_offset": 0, 00:07:19.239 "data_size": 65536 00:07:19.239 }, 00:07:19.239 { 00:07:19.239 "name": "BaseBdev2", 00:07:19.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.239 "is_configured": false, 00:07:19.239 "data_offset": 0, 00:07:19.239 "data_size": 0 00:07:19.239 } 00:07:19.239 ] 00:07:19.239 }' 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.239 14:06:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.498 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:19.498 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.498 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.498 [2024-09-30 14:06:24.062821] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:19.498 [2024-09-30 14:06:24.062935] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:19.499 [2024-09-30 14:06:24.062984] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:19.499 [2024-09-30 14:06:24.063355] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:19.499 [2024-09-30 14:06:24.063582] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:19.499 [2024-09-30 14:06:24.063628] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:19.499 [2024-09-30 14:06:24.063903] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:19.499 BaseBdev2 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.499 [ 00:07:19.499 { 00:07:19.499 "name": "BaseBdev2", 00:07:19.499 "aliases": [ 00:07:19.499 "55d6936b-acc4-4eef-a4ad-c1b8586ec881" 00:07:19.499 ], 00:07:19.499 "product_name": "Malloc disk", 00:07:19.499 "block_size": 512, 00:07:19.499 "num_blocks": 65536, 00:07:19.499 "uuid": "55d6936b-acc4-4eef-a4ad-c1b8586ec881", 00:07:19.499 "assigned_rate_limits": { 00:07:19.499 "rw_ios_per_sec": 0, 00:07:19.499 "rw_mbytes_per_sec": 0, 00:07:19.499 "r_mbytes_per_sec": 0, 00:07:19.499 "w_mbytes_per_sec": 0 00:07:19.499 }, 00:07:19.499 "claimed": true, 00:07:19.499 "claim_type": "exclusive_write", 00:07:19.499 "zoned": false, 00:07:19.499 "supported_io_types": { 00:07:19.499 "read": true, 00:07:19.499 "write": true, 00:07:19.499 "unmap": true, 00:07:19.499 "flush": true, 00:07:19.499 "reset": true, 00:07:19.499 "nvme_admin": false, 00:07:19.499 "nvme_io": false, 00:07:19.499 "nvme_io_md": false, 00:07:19.499 "write_zeroes": true, 00:07:19.499 "zcopy": true, 00:07:19.499 "get_zone_info": false, 00:07:19.499 "zone_management": false, 00:07:19.499 "zone_append": false, 00:07:19.499 "compare": false, 00:07:19.499 "compare_and_write": false, 00:07:19.499 "abort": true, 00:07:19.499 "seek_hole": false, 00:07:19.499 "seek_data": false, 00:07:19.499 "copy": true, 00:07:19.499 "nvme_iov_md": false 00:07:19.499 }, 00:07:19.499 "memory_domains": [ 00:07:19.499 { 00:07:19.499 "dma_device_id": "system", 00:07:19.499 "dma_device_type": 1 00:07:19.499 }, 00:07:19.499 { 00:07:19.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:19.499 "dma_device_type": 2 00:07:19.499 } 00:07:19.499 ], 00:07:19.499 "driver_specific": {} 00:07:19.499 } 00:07:19.499 ] 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.499 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.759 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.759 "name": "Existed_Raid", 00:07:19.759 "uuid": "db59effd-adc3-41db-bdec-b461e67b4dcb", 00:07:19.759 "strip_size_kb": 64, 00:07:19.759 "state": "online", 00:07:19.759 "raid_level": "concat", 00:07:19.759 "superblock": false, 00:07:19.759 "num_base_bdevs": 2, 00:07:19.759 "num_base_bdevs_discovered": 2, 00:07:19.759 "num_base_bdevs_operational": 2, 00:07:19.759 "base_bdevs_list": [ 00:07:19.759 { 00:07:19.759 "name": "BaseBdev1", 00:07:19.759 "uuid": "07e905ff-b20b-4941-93e7-a6b196c00b40", 00:07:19.759 "is_configured": true, 00:07:19.759 "data_offset": 0, 00:07:19.759 "data_size": 65536 00:07:19.759 }, 00:07:19.759 { 00:07:19.759 "name": "BaseBdev2", 00:07:19.759 "uuid": "55d6936b-acc4-4eef-a4ad-c1b8586ec881", 00:07:19.759 "is_configured": true, 00:07:19.759 "data_offset": 0, 00:07:19.759 "data_size": 65536 00:07:19.759 } 00:07:19.759 ] 00:07:19.759 }' 00:07:19.759 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.759 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.017 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:20.018 [2024-09-30 14:06:24.506860] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:20.018 "name": "Existed_Raid", 00:07:20.018 "aliases": [ 00:07:20.018 "db59effd-adc3-41db-bdec-b461e67b4dcb" 00:07:20.018 ], 00:07:20.018 "product_name": "Raid Volume", 00:07:20.018 "block_size": 512, 00:07:20.018 "num_blocks": 131072, 00:07:20.018 "uuid": "db59effd-adc3-41db-bdec-b461e67b4dcb", 00:07:20.018 "assigned_rate_limits": { 00:07:20.018 "rw_ios_per_sec": 0, 00:07:20.018 "rw_mbytes_per_sec": 0, 00:07:20.018 "r_mbytes_per_sec": 0, 00:07:20.018 "w_mbytes_per_sec": 0 00:07:20.018 }, 00:07:20.018 "claimed": false, 00:07:20.018 "zoned": false, 00:07:20.018 "supported_io_types": { 00:07:20.018 "read": true, 00:07:20.018 "write": true, 00:07:20.018 "unmap": true, 00:07:20.018 "flush": true, 00:07:20.018 "reset": true, 00:07:20.018 "nvme_admin": false, 00:07:20.018 "nvme_io": false, 00:07:20.018 "nvme_io_md": false, 00:07:20.018 "write_zeroes": true, 00:07:20.018 "zcopy": false, 00:07:20.018 "get_zone_info": false, 00:07:20.018 "zone_management": false, 00:07:20.018 "zone_append": false, 00:07:20.018 "compare": false, 00:07:20.018 "compare_and_write": false, 00:07:20.018 "abort": false, 00:07:20.018 "seek_hole": false, 00:07:20.018 "seek_data": false, 00:07:20.018 "copy": false, 00:07:20.018 "nvme_iov_md": false 00:07:20.018 }, 00:07:20.018 "memory_domains": [ 00:07:20.018 { 00:07:20.018 "dma_device_id": "system", 00:07:20.018 "dma_device_type": 1 00:07:20.018 }, 00:07:20.018 { 00:07:20.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.018 "dma_device_type": 2 00:07:20.018 }, 00:07:20.018 { 00:07:20.018 "dma_device_id": "system", 00:07:20.018 "dma_device_type": 1 00:07:20.018 }, 00:07:20.018 { 00:07:20.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.018 "dma_device_type": 2 00:07:20.018 } 00:07:20.018 ], 00:07:20.018 "driver_specific": { 00:07:20.018 "raid": { 00:07:20.018 "uuid": "db59effd-adc3-41db-bdec-b461e67b4dcb", 00:07:20.018 "strip_size_kb": 64, 00:07:20.018 "state": "online", 00:07:20.018 "raid_level": "concat", 00:07:20.018 "superblock": false, 00:07:20.018 "num_base_bdevs": 2, 00:07:20.018 "num_base_bdevs_discovered": 2, 00:07:20.018 "num_base_bdevs_operational": 2, 00:07:20.018 "base_bdevs_list": [ 00:07:20.018 { 00:07:20.018 "name": "BaseBdev1", 00:07:20.018 "uuid": "07e905ff-b20b-4941-93e7-a6b196c00b40", 00:07:20.018 "is_configured": true, 00:07:20.018 "data_offset": 0, 00:07:20.018 "data_size": 65536 00:07:20.018 }, 00:07:20.018 { 00:07:20.018 "name": "BaseBdev2", 00:07:20.018 "uuid": "55d6936b-acc4-4eef-a4ad-c1b8586ec881", 00:07:20.018 "is_configured": true, 00:07:20.018 "data_offset": 0, 00:07:20.018 "data_size": 65536 00:07:20.018 } 00:07:20.018 ] 00:07:20.018 } 00:07:20.018 } 00:07:20.018 }' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:20.018 BaseBdev2' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.018 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.277 [2024-09-30 14:06:24.710641] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:20.277 [2024-09-30 14:06:24.710704] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:20.277 [2024-09-30 14:06:24.710773] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.277 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.278 "name": "Existed_Raid", 00:07:20.278 "uuid": "db59effd-adc3-41db-bdec-b461e67b4dcb", 00:07:20.278 "strip_size_kb": 64, 00:07:20.278 "state": "offline", 00:07:20.278 "raid_level": "concat", 00:07:20.278 "superblock": false, 00:07:20.278 "num_base_bdevs": 2, 00:07:20.278 "num_base_bdevs_discovered": 1, 00:07:20.278 "num_base_bdevs_operational": 1, 00:07:20.278 "base_bdevs_list": [ 00:07:20.278 { 00:07:20.278 "name": null, 00:07:20.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.278 "is_configured": false, 00:07:20.278 "data_offset": 0, 00:07:20.278 "data_size": 65536 00:07:20.278 }, 00:07:20.278 { 00:07:20.278 "name": "BaseBdev2", 00:07:20.278 "uuid": "55d6936b-acc4-4eef-a4ad-c1b8586ec881", 00:07:20.278 "is_configured": true, 00:07:20.278 "data_offset": 0, 00:07:20.278 "data_size": 65536 00:07:20.278 } 00:07:20.278 ] 00:07:20.278 }' 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.278 14:06:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.538 [2024-09-30 14:06:25.125072] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:20.538 [2024-09-30 14:06:25.125141] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74094 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 74094 ']' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 74094 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:20.538 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74094 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74094' 00:07:20.798 killing process with pid 74094 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 74094 00:07:20.798 [2024-09-30 14:06:25.200039] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 74094 00:07:20.798 [2024-09-30 14:06:25.201001] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:20.798 ************************************ 00:07:20.798 END TEST raid_state_function_test 00:07:20.798 ************************************ 00:07:20.798 00:07:20.798 real 0m3.620s 00:07:20.798 user 0m5.642s 00:07:20.798 sys 0m0.718s 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:20.798 14:06:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.059 14:06:25 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:21.059 14:06:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:21.059 14:06:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:21.059 14:06:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:21.059 ************************************ 00:07:21.059 START TEST raid_state_function_test_sb 00:07:21.059 ************************************ 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 true 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74336 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:21.059 Process raid pid: 74336 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74336' 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74336 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 74336 ']' 00:07:21.059 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:21.059 14:06:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.059 [2024-09-30 14:06:25.603997] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:21.059 [2024-09-30 14:06:25.604134] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:21.319 [2024-09-30 14:06:25.735688] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:21.319 [2024-09-30 14:06:25.765718] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:21.319 [2024-09-30 14:06:25.810151] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:21.319 [2024-09-30 14:06:25.851430] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:21.319 [2024-09-30 14:06:25.851465] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.888 [2024-09-30 14:06:26.424728] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:21.888 [2024-09-30 14:06:26.424776] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:21.888 [2024-09-30 14:06:26.424789] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:21.888 [2024-09-30 14:06:26.424797] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.888 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.888 "name": "Existed_Raid", 00:07:21.888 "uuid": "f11ac76a-1007-4026-8935-833ebc48d4f3", 00:07:21.888 "strip_size_kb": 64, 00:07:21.888 "state": "configuring", 00:07:21.888 "raid_level": "concat", 00:07:21.888 "superblock": true, 00:07:21.888 "num_base_bdevs": 2, 00:07:21.888 "num_base_bdevs_discovered": 0, 00:07:21.888 "num_base_bdevs_operational": 2, 00:07:21.888 "base_bdevs_list": [ 00:07:21.888 { 00:07:21.888 "name": "BaseBdev1", 00:07:21.888 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.888 "is_configured": false, 00:07:21.888 "data_offset": 0, 00:07:21.888 "data_size": 0 00:07:21.888 }, 00:07:21.888 { 00:07:21.888 "name": "BaseBdev2", 00:07:21.888 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.888 "is_configured": false, 00:07:21.888 "data_offset": 0, 00:07:21.888 "data_size": 0 00:07:21.888 } 00:07:21.888 ] 00:07:21.889 }' 00:07:21.889 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.889 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 [2024-09-30 14:06:26.827916] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:22.457 [2024-09-30 14:06:26.827999] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 [2024-09-30 14:06:26.835935] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:22.457 [2024-09-30 14:06:26.836006] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:22.457 [2024-09-30 14:06:26.836034] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:22.457 [2024-09-30 14:06:26.836056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 [2024-09-30 14:06:26.852529] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:22.457 BaseBdev1 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.457 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.457 [ 00:07:22.457 { 00:07:22.457 "name": "BaseBdev1", 00:07:22.457 "aliases": [ 00:07:22.457 "dafd7b71-4097-4856-9eb8-03a2231ee01b" 00:07:22.457 ], 00:07:22.457 "product_name": "Malloc disk", 00:07:22.457 "block_size": 512, 00:07:22.458 "num_blocks": 65536, 00:07:22.458 "uuid": "dafd7b71-4097-4856-9eb8-03a2231ee01b", 00:07:22.458 "assigned_rate_limits": { 00:07:22.458 "rw_ios_per_sec": 0, 00:07:22.458 "rw_mbytes_per_sec": 0, 00:07:22.458 "r_mbytes_per_sec": 0, 00:07:22.458 "w_mbytes_per_sec": 0 00:07:22.458 }, 00:07:22.458 "claimed": true, 00:07:22.458 "claim_type": "exclusive_write", 00:07:22.458 "zoned": false, 00:07:22.458 "supported_io_types": { 00:07:22.458 "read": true, 00:07:22.458 "write": true, 00:07:22.458 "unmap": true, 00:07:22.458 "flush": true, 00:07:22.458 "reset": true, 00:07:22.458 "nvme_admin": false, 00:07:22.458 "nvme_io": false, 00:07:22.458 "nvme_io_md": false, 00:07:22.458 "write_zeroes": true, 00:07:22.458 "zcopy": true, 00:07:22.458 "get_zone_info": false, 00:07:22.458 "zone_management": false, 00:07:22.458 "zone_append": false, 00:07:22.458 "compare": false, 00:07:22.458 "compare_and_write": false, 00:07:22.458 "abort": true, 00:07:22.458 "seek_hole": false, 00:07:22.458 "seek_data": false, 00:07:22.458 "copy": true, 00:07:22.458 "nvme_iov_md": false 00:07:22.458 }, 00:07:22.458 "memory_domains": [ 00:07:22.458 { 00:07:22.458 "dma_device_id": "system", 00:07:22.458 "dma_device_type": 1 00:07:22.458 }, 00:07:22.458 { 00:07:22.458 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.458 "dma_device_type": 2 00:07:22.458 } 00:07:22.458 ], 00:07:22.458 "driver_specific": {} 00:07:22.458 } 00:07:22.458 ] 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.458 "name": "Existed_Raid", 00:07:22.458 "uuid": "a96c0c9e-9fea-4238-8463-5f691716b876", 00:07:22.458 "strip_size_kb": 64, 00:07:22.458 "state": "configuring", 00:07:22.458 "raid_level": "concat", 00:07:22.458 "superblock": true, 00:07:22.458 "num_base_bdevs": 2, 00:07:22.458 "num_base_bdevs_discovered": 1, 00:07:22.458 "num_base_bdevs_operational": 2, 00:07:22.458 "base_bdevs_list": [ 00:07:22.458 { 00:07:22.458 "name": "BaseBdev1", 00:07:22.458 "uuid": "dafd7b71-4097-4856-9eb8-03a2231ee01b", 00:07:22.458 "is_configured": true, 00:07:22.458 "data_offset": 2048, 00:07:22.458 "data_size": 63488 00:07:22.458 }, 00:07:22.458 { 00:07:22.458 "name": "BaseBdev2", 00:07:22.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.458 "is_configured": false, 00:07:22.458 "data_offset": 0, 00:07:22.458 "data_size": 0 00:07:22.458 } 00:07:22.458 ] 00:07:22.458 }' 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.458 14:06:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.718 [2024-09-30 14:06:27.283809] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:22.718 [2024-09-30 14:06:27.283858] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.718 [2024-09-30 14:06:27.295812] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:22.718 [2024-09-30 14:06:27.297559] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:22.718 [2024-09-30 14:06:27.297596] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.718 "name": "Existed_Raid", 00:07:22.718 "uuid": "00f2bb42-f580-4bb6-bfa9-87043a446fb4", 00:07:22.718 "strip_size_kb": 64, 00:07:22.718 "state": "configuring", 00:07:22.718 "raid_level": "concat", 00:07:22.718 "superblock": true, 00:07:22.718 "num_base_bdevs": 2, 00:07:22.718 "num_base_bdevs_discovered": 1, 00:07:22.718 "num_base_bdevs_operational": 2, 00:07:22.718 "base_bdevs_list": [ 00:07:22.718 { 00:07:22.718 "name": "BaseBdev1", 00:07:22.718 "uuid": "dafd7b71-4097-4856-9eb8-03a2231ee01b", 00:07:22.718 "is_configured": true, 00:07:22.718 "data_offset": 2048, 00:07:22.718 "data_size": 63488 00:07:22.718 }, 00:07:22.718 { 00:07:22.718 "name": "BaseBdev2", 00:07:22.718 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.718 "is_configured": false, 00:07:22.718 "data_offset": 0, 00:07:22.718 "data_size": 0 00:07:22.718 } 00:07:22.718 ] 00:07:22.718 }' 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.718 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.287 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.288 [2024-09-30 14:06:27.748308] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:23.288 [2024-09-30 14:06:27.749008] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:23.288 [2024-09-30 14:06:27.749186] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:23.288 BaseBdev2 00:07:23.288 [2024-09-30 14:06:27.750288] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:23.288 [2024-09-30 14:06:27.750985] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:23.288 [2024-09-30 14:06:27.751141] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:23.288 [2024-09-30 14:06:27.751832] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.288 [ 00:07:23.288 { 00:07:23.288 "name": "BaseBdev2", 00:07:23.288 "aliases": [ 00:07:23.288 "7abe028f-4971-4fa7-aca0-a833f98f1b94" 00:07:23.288 ], 00:07:23.288 "product_name": "Malloc disk", 00:07:23.288 "block_size": 512, 00:07:23.288 "num_blocks": 65536, 00:07:23.288 "uuid": "7abe028f-4971-4fa7-aca0-a833f98f1b94", 00:07:23.288 "assigned_rate_limits": { 00:07:23.288 "rw_ios_per_sec": 0, 00:07:23.288 "rw_mbytes_per_sec": 0, 00:07:23.288 "r_mbytes_per_sec": 0, 00:07:23.288 "w_mbytes_per_sec": 0 00:07:23.288 }, 00:07:23.288 "claimed": true, 00:07:23.288 "claim_type": "exclusive_write", 00:07:23.288 "zoned": false, 00:07:23.288 "supported_io_types": { 00:07:23.288 "read": true, 00:07:23.288 "write": true, 00:07:23.288 "unmap": true, 00:07:23.288 "flush": true, 00:07:23.288 "reset": true, 00:07:23.288 "nvme_admin": false, 00:07:23.288 "nvme_io": false, 00:07:23.288 "nvme_io_md": false, 00:07:23.288 "write_zeroes": true, 00:07:23.288 "zcopy": true, 00:07:23.288 "get_zone_info": false, 00:07:23.288 "zone_management": false, 00:07:23.288 "zone_append": false, 00:07:23.288 "compare": false, 00:07:23.288 "compare_and_write": false, 00:07:23.288 "abort": true, 00:07:23.288 "seek_hole": false, 00:07:23.288 "seek_data": false, 00:07:23.288 "copy": true, 00:07:23.288 "nvme_iov_md": false 00:07:23.288 }, 00:07:23.288 "memory_domains": [ 00:07:23.288 { 00:07:23.288 "dma_device_id": "system", 00:07:23.288 "dma_device_type": 1 00:07:23.288 }, 00:07:23.288 { 00:07:23.288 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.288 "dma_device_type": 2 00:07:23.288 } 00:07:23.288 ], 00:07:23.288 "driver_specific": {} 00:07:23.288 } 00:07:23.288 ] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.288 "name": "Existed_Raid", 00:07:23.288 "uuid": "00f2bb42-f580-4bb6-bfa9-87043a446fb4", 00:07:23.288 "strip_size_kb": 64, 00:07:23.288 "state": "online", 00:07:23.288 "raid_level": "concat", 00:07:23.288 "superblock": true, 00:07:23.288 "num_base_bdevs": 2, 00:07:23.288 "num_base_bdevs_discovered": 2, 00:07:23.288 "num_base_bdevs_operational": 2, 00:07:23.288 "base_bdevs_list": [ 00:07:23.288 { 00:07:23.288 "name": "BaseBdev1", 00:07:23.288 "uuid": "dafd7b71-4097-4856-9eb8-03a2231ee01b", 00:07:23.288 "is_configured": true, 00:07:23.288 "data_offset": 2048, 00:07:23.288 "data_size": 63488 00:07:23.288 }, 00:07:23.288 { 00:07:23.288 "name": "BaseBdev2", 00:07:23.288 "uuid": "7abe028f-4971-4fa7-aca0-a833f98f1b94", 00:07:23.288 "is_configured": true, 00:07:23.288 "data_offset": 2048, 00:07:23.288 "data_size": 63488 00:07:23.288 } 00:07:23.288 ] 00:07:23.288 }' 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.288 14:06:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.857 [2024-09-30 14:06:28.219727] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:23.857 "name": "Existed_Raid", 00:07:23.857 "aliases": [ 00:07:23.857 "00f2bb42-f580-4bb6-bfa9-87043a446fb4" 00:07:23.857 ], 00:07:23.857 "product_name": "Raid Volume", 00:07:23.857 "block_size": 512, 00:07:23.857 "num_blocks": 126976, 00:07:23.857 "uuid": "00f2bb42-f580-4bb6-bfa9-87043a446fb4", 00:07:23.857 "assigned_rate_limits": { 00:07:23.857 "rw_ios_per_sec": 0, 00:07:23.857 "rw_mbytes_per_sec": 0, 00:07:23.857 "r_mbytes_per_sec": 0, 00:07:23.857 "w_mbytes_per_sec": 0 00:07:23.857 }, 00:07:23.857 "claimed": false, 00:07:23.857 "zoned": false, 00:07:23.857 "supported_io_types": { 00:07:23.857 "read": true, 00:07:23.857 "write": true, 00:07:23.857 "unmap": true, 00:07:23.857 "flush": true, 00:07:23.857 "reset": true, 00:07:23.857 "nvme_admin": false, 00:07:23.857 "nvme_io": false, 00:07:23.857 "nvme_io_md": false, 00:07:23.857 "write_zeroes": true, 00:07:23.857 "zcopy": false, 00:07:23.857 "get_zone_info": false, 00:07:23.857 "zone_management": false, 00:07:23.857 "zone_append": false, 00:07:23.857 "compare": false, 00:07:23.857 "compare_and_write": false, 00:07:23.857 "abort": false, 00:07:23.857 "seek_hole": false, 00:07:23.857 "seek_data": false, 00:07:23.857 "copy": false, 00:07:23.857 "nvme_iov_md": false 00:07:23.857 }, 00:07:23.857 "memory_domains": [ 00:07:23.857 { 00:07:23.857 "dma_device_id": "system", 00:07:23.857 "dma_device_type": 1 00:07:23.857 }, 00:07:23.857 { 00:07:23.857 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.857 "dma_device_type": 2 00:07:23.857 }, 00:07:23.857 { 00:07:23.857 "dma_device_id": "system", 00:07:23.857 "dma_device_type": 1 00:07:23.857 }, 00:07:23.857 { 00:07:23.857 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.857 "dma_device_type": 2 00:07:23.857 } 00:07:23.857 ], 00:07:23.857 "driver_specific": { 00:07:23.857 "raid": { 00:07:23.857 "uuid": "00f2bb42-f580-4bb6-bfa9-87043a446fb4", 00:07:23.857 "strip_size_kb": 64, 00:07:23.857 "state": "online", 00:07:23.857 "raid_level": "concat", 00:07:23.857 "superblock": true, 00:07:23.857 "num_base_bdevs": 2, 00:07:23.857 "num_base_bdevs_discovered": 2, 00:07:23.857 "num_base_bdevs_operational": 2, 00:07:23.857 "base_bdevs_list": [ 00:07:23.857 { 00:07:23.857 "name": "BaseBdev1", 00:07:23.857 "uuid": "dafd7b71-4097-4856-9eb8-03a2231ee01b", 00:07:23.857 "is_configured": true, 00:07:23.857 "data_offset": 2048, 00:07:23.857 "data_size": 63488 00:07:23.857 }, 00:07:23.857 { 00:07:23.857 "name": "BaseBdev2", 00:07:23.857 "uuid": "7abe028f-4971-4fa7-aca0-a833f98f1b94", 00:07:23.857 "is_configured": true, 00:07:23.857 "data_offset": 2048, 00:07:23.857 "data_size": 63488 00:07:23.857 } 00:07:23.857 ] 00:07:23.857 } 00:07:23.857 } 00:07:23.857 }' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:23.857 BaseBdev2' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.857 [2024-09-30 14:06:28.399215] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:23.857 [2024-09-30 14:06:28.399280] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:23.857 [2024-09-30 14:06:28.399332] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:23.857 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.858 "name": "Existed_Raid", 00:07:23.858 "uuid": "00f2bb42-f580-4bb6-bfa9-87043a446fb4", 00:07:23.858 "strip_size_kb": 64, 00:07:23.858 "state": "offline", 00:07:23.858 "raid_level": "concat", 00:07:23.858 "superblock": true, 00:07:23.858 "num_base_bdevs": 2, 00:07:23.858 "num_base_bdevs_discovered": 1, 00:07:23.858 "num_base_bdevs_operational": 1, 00:07:23.858 "base_bdevs_list": [ 00:07:23.858 { 00:07:23.858 "name": null, 00:07:23.858 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:23.858 "is_configured": false, 00:07:23.858 "data_offset": 0, 00:07:23.858 "data_size": 63488 00:07:23.858 }, 00:07:23.858 { 00:07:23.858 "name": "BaseBdev2", 00:07:23.858 "uuid": "7abe028f-4971-4fa7-aca0-a833f98f1b94", 00:07:23.858 "is_configured": true, 00:07:23.858 "data_offset": 2048, 00:07:23.858 "data_size": 63488 00:07:23.858 } 00:07:23.858 ] 00:07:23.858 }' 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.858 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.428 [2024-09-30 14:06:28.893604] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:24.428 [2024-09-30 14:06:28.893693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74336 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 74336 ']' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 74336 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:24.428 14:06:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74336 00:07:24.428 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:24.428 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:24.428 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74336' 00:07:24.428 killing process with pid 74336 00:07:24.428 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 74336 00:07:24.428 [2024-09-30 14:06:29.008845] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:24.428 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 74336 00:07:24.428 [2024-09-30 14:06:29.009804] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:24.687 14:06:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:24.687 00:07:24.688 real 0m3.740s 00:07:24.688 user 0m5.847s 00:07:24.688 sys 0m0.742s 00:07:24.688 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:24.688 ************************************ 00:07:24.688 END TEST raid_state_function_test_sb 00:07:24.688 ************************************ 00:07:24.688 14:06:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.688 14:06:29 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:24.688 14:06:29 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:24.688 14:06:29 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:24.688 14:06:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:24.688 ************************************ 00:07:24.688 START TEST raid_superblock_test 00:07:24.688 ************************************ 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 2 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74576 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74576 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 74576 ']' 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:24.688 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:24.688 14:06:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.948 [2024-09-30 14:06:29.424266] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:24.948 [2024-09-30 14:06:29.424521] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74576 ] 00:07:24.948 [2024-09-30 14:06:29.563780] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:24.948 [2024-09-30 14:06:29.590244] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:25.207 [2024-09-30 14:06:29.635314] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:25.207 [2024-09-30 14:06:29.677585] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.207 [2024-09-30 14:06:29.677692] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 malloc1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 [2024-09-30 14:06:30.239597] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:25.777 [2024-09-30 14:06:30.239723] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.777 [2024-09-30 14:06:30.239782] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:25.777 [2024-09-30 14:06:30.239841] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.777 [2024-09-30 14:06:30.241900] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.777 [2024-09-30 14:06:30.241968] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:25.777 pt1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 malloc2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 [2024-09-30 14:06:30.287524] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:25.777 [2024-09-30 14:06:30.287626] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.777 [2024-09-30 14:06:30.287668] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:25.777 [2024-09-30 14:06:30.287691] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.777 [2024-09-30 14:06:30.292179] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.777 [2024-09-30 14:06:30.292229] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:25.777 pt2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 [2024-09-30 14:06:30.300522] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:25.777 [2024-09-30 14:06:30.302809] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:25.777 [2024-09-30 14:06:30.302964] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:07:25.777 [2024-09-30 14:06:30.302978] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:25.777 [2024-09-30 14:06:30.303271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:25.777 [2024-09-30 14:06:30.303406] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:07:25.777 [2024-09-30 14:06:30.303420] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:07:25.777 [2024-09-30 14:06:30.303578] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.777 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.777 "name": "raid_bdev1", 00:07:25.777 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:25.778 "strip_size_kb": 64, 00:07:25.778 "state": "online", 00:07:25.778 "raid_level": "concat", 00:07:25.778 "superblock": true, 00:07:25.778 "num_base_bdevs": 2, 00:07:25.778 "num_base_bdevs_discovered": 2, 00:07:25.778 "num_base_bdevs_operational": 2, 00:07:25.778 "base_bdevs_list": [ 00:07:25.778 { 00:07:25.778 "name": "pt1", 00:07:25.778 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.778 "is_configured": true, 00:07:25.778 "data_offset": 2048, 00:07:25.778 "data_size": 63488 00:07:25.778 }, 00:07:25.778 { 00:07:25.778 "name": "pt2", 00:07:25.778 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.778 "is_configured": true, 00:07:25.778 "data_offset": 2048, 00:07:25.778 "data_size": 63488 00:07:25.778 } 00:07:25.778 ] 00:07:25.778 }' 00:07:25.778 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.778 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.347 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.347 [2024-09-30 14:06:30.739897] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:26.348 "name": "raid_bdev1", 00:07:26.348 "aliases": [ 00:07:26.348 "c326b647-184e-4706-a9c3-63b2fcb816c8" 00:07:26.348 ], 00:07:26.348 "product_name": "Raid Volume", 00:07:26.348 "block_size": 512, 00:07:26.348 "num_blocks": 126976, 00:07:26.348 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:26.348 "assigned_rate_limits": { 00:07:26.348 "rw_ios_per_sec": 0, 00:07:26.348 "rw_mbytes_per_sec": 0, 00:07:26.348 "r_mbytes_per_sec": 0, 00:07:26.348 "w_mbytes_per_sec": 0 00:07:26.348 }, 00:07:26.348 "claimed": false, 00:07:26.348 "zoned": false, 00:07:26.348 "supported_io_types": { 00:07:26.348 "read": true, 00:07:26.348 "write": true, 00:07:26.348 "unmap": true, 00:07:26.348 "flush": true, 00:07:26.348 "reset": true, 00:07:26.348 "nvme_admin": false, 00:07:26.348 "nvme_io": false, 00:07:26.348 "nvme_io_md": false, 00:07:26.348 "write_zeroes": true, 00:07:26.348 "zcopy": false, 00:07:26.348 "get_zone_info": false, 00:07:26.348 "zone_management": false, 00:07:26.348 "zone_append": false, 00:07:26.348 "compare": false, 00:07:26.348 "compare_and_write": false, 00:07:26.348 "abort": false, 00:07:26.348 "seek_hole": false, 00:07:26.348 "seek_data": false, 00:07:26.348 "copy": false, 00:07:26.348 "nvme_iov_md": false 00:07:26.348 }, 00:07:26.348 "memory_domains": [ 00:07:26.348 { 00:07:26.348 "dma_device_id": "system", 00:07:26.348 "dma_device_type": 1 00:07:26.348 }, 00:07:26.348 { 00:07:26.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.348 "dma_device_type": 2 00:07:26.348 }, 00:07:26.348 { 00:07:26.348 "dma_device_id": "system", 00:07:26.348 "dma_device_type": 1 00:07:26.348 }, 00:07:26.348 { 00:07:26.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.348 "dma_device_type": 2 00:07:26.348 } 00:07:26.348 ], 00:07:26.348 "driver_specific": { 00:07:26.348 "raid": { 00:07:26.348 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:26.348 "strip_size_kb": 64, 00:07:26.348 "state": "online", 00:07:26.348 "raid_level": "concat", 00:07:26.348 "superblock": true, 00:07:26.348 "num_base_bdevs": 2, 00:07:26.348 "num_base_bdevs_discovered": 2, 00:07:26.348 "num_base_bdevs_operational": 2, 00:07:26.348 "base_bdevs_list": [ 00:07:26.348 { 00:07:26.348 "name": "pt1", 00:07:26.348 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.348 "is_configured": true, 00:07:26.348 "data_offset": 2048, 00:07:26.348 "data_size": 63488 00:07:26.348 }, 00:07:26.348 { 00:07:26.348 "name": "pt2", 00:07:26.348 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.348 "is_configured": true, 00:07:26.348 "data_offset": 2048, 00:07:26.348 "data_size": 63488 00:07:26.348 } 00:07:26.348 ] 00:07:26.348 } 00:07:26.348 } 00:07:26.348 }' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:26.348 pt2' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.348 [2024-09-30 14:06:30.947491] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c326b647-184e-4706-a9c3-63b2fcb816c8 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z c326b647-184e-4706-a9c3-63b2fcb816c8 ']' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.348 [2024-09-30 14:06:30.987178] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:26.348 [2024-09-30 14:06:30.987239] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:26.348 [2024-09-30 14:06:30.987305] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.348 [2024-09-30 14:06:30.987356] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.348 [2024-09-30 14:06:30.987374] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:26.348 14:06:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.607 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.607 [2024-09-30 14:06:31.122979] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:26.607 [2024-09-30 14:06:31.124858] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:26.607 [2024-09-30 14:06:31.124954] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:26.607 [2024-09-30 14:06:31.125046] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:26.607 [2024-09-30 14:06:31.125084] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:26.607 [2024-09-30 14:06:31.125106] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:07:26.607 request: 00:07:26.607 { 00:07:26.607 "name": "raid_bdev1", 00:07:26.607 "raid_level": "concat", 00:07:26.607 "base_bdevs": [ 00:07:26.607 "malloc1", 00:07:26.607 "malloc2" 00:07:26.607 ], 00:07:26.607 "strip_size_kb": 64, 00:07:26.607 "superblock": false, 00:07:26.607 "method": "bdev_raid_create", 00:07:26.608 "req_id": 1 00:07:26.608 } 00:07:26.608 Got JSON-RPC error response 00:07:26.608 response: 00:07:26.608 { 00:07:26.608 "code": -17, 00:07:26.608 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:26.608 } 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.608 [2024-09-30 14:06:31.178865] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:26.608 [2024-09-30 14:06:31.178947] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:26.608 [2024-09-30 14:06:31.178979] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:26.608 [2024-09-30 14:06:31.179013] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:26.608 [2024-09-30 14:06:31.181039] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:26.608 [2024-09-30 14:06:31.181106] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:26.608 [2024-09-30 14:06:31.181203] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:26.608 [2024-09-30 14:06:31.181278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:26.608 pt1 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:26.608 "name": "raid_bdev1", 00:07:26.608 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:26.608 "strip_size_kb": 64, 00:07:26.608 "state": "configuring", 00:07:26.608 "raid_level": "concat", 00:07:26.608 "superblock": true, 00:07:26.608 "num_base_bdevs": 2, 00:07:26.608 "num_base_bdevs_discovered": 1, 00:07:26.608 "num_base_bdevs_operational": 2, 00:07:26.608 "base_bdevs_list": [ 00:07:26.608 { 00:07:26.608 "name": "pt1", 00:07:26.608 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.608 "is_configured": true, 00:07:26.608 "data_offset": 2048, 00:07:26.608 "data_size": 63488 00:07:26.608 }, 00:07:26.608 { 00:07:26.608 "name": null, 00:07:26.608 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.608 "is_configured": false, 00:07:26.608 "data_offset": 2048, 00:07:26.608 "data_size": 63488 00:07:26.608 } 00:07:26.608 ] 00:07:26.608 }' 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:26.608 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.174 [2024-09-30 14:06:31.662042] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:27.174 [2024-09-30 14:06:31.662100] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.174 [2024-09-30 14:06:31.662120] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:27.174 [2024-09-30 14:06:31.662129] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.174 [2024-09-30 14:06:31.662518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.174 [2024-09-30 14:06:31.662538] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:27.174 [2024-09-30 14:06:31.662604] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:27.174 [2024-09-30 14:06:31.662625] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:27.174 [2024-09-30 14:06:31.662705] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:27.174 [2024-09-30 14:06:31.662716] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:27.174 [2024-09-30 14:06:31.662932] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:27.174 [2024-09-30 14:06:31.663053] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:27.174 [2024-09-30 14:06:31.663063] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:27.174 [2024-09-30 14:06:31.663153] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.174 pt2 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:27.174 "name": "raid_bdev1", 00:07:27.174 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:27.174 "strip_size_kb": 64, 00:07:27.174 "state": "online", 00:07:27.174 "raid_level": "concat", 00:07:27.174 "superblock": true, 00:07:27.174 "num_base_bdevs": 2, 00:07:27.174 "num_base_bdevs_discovered": 2, 00:07:27.174 "num_base_bdevs_operational": 2, 00:07:27.174 "base_bdevs_list": [ 00:07:27.174 { 00:07:27.174 "name": "pt1", 00:07:27.174 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:27.174 "is_configured": true, 00:07:27.174 "data_offset": 2048, 00:07:27.174 "data_size": 63488 00:07:27.174 }, 00:07:27.174 { 00:07:27.174 "name": "pt2", 00:07:27.174 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:27.174 "is_configured": true, 00:07:27.174 "data_offset": 2048, 00:07:27.174 "data_size": 63488 00:07:27.174 } 00:07:27.174 ] 00:07:27.174 }' 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:27.174 14:06:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.742 [2024-09-30 14:06:32.145425] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:27.742 "name": "raid_bdev1", 00:07:27.742 "aliases": [ 00:07:27.742 "c326b647-184e-4706-a9c3-63b2fcb816c8" 00:07:27.742 ], 00:07:27.742 "product_name": "Raid Volume", 00:07:27.742 "block_size": 512, 00:07:27.742 "num_blocks": 126976, 00:07:27.742 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:27.742 "assigned_rate_limits": { 00:07:27.742 "rw_ios_per_sec": 0, 00:07:27.742 "rw_mbytes_per_sec": 0, 00:07:27.742 "r_mbytes_per_sec": 0, 00:07:27.742 "w_mbytes_per_sec": 0 00:07:27.742 }, 00:07:27.742 "claimed": false, 00:07:27.742 "zoned": false, 00:07:27.742 "supported_io_types": { 00:07:27.742 "read": true, 00:07:27.742 "write": true, 00:07:27.742 "unmap": true, 00:07:27.742 "flush": true, 00:07:27.742 "reset": true, 00:07:27.742 "nvme_admin": false, 00:07:27.742 "nvme_io": false, 00:07:27.742 "nvme_io_md": false, 00:07:27.742 "write_zeroes": true, 00:07:27.742 "zcopy": false, 00:07:27.742 "get_zone_info": false, 00:07:27.742 "zone_management": false, 00:07:27.742 "zone_append": false, 00:07:27.742 "compare": false, 00:07:27.742 "compare_and_write": false, 00:07:27.742 "abort": false, 00:07:27.742 "seek_hole": false, 00:07:27.742 "seek_data": false, 00:07:27.742 "copy": false, 00:07:27.742 "nvme_iov_md": false 00:07:27.742 }, 00:07:27.742 "memory_domains": [ 00:07:27.742 { 00:07:27.742 "dma_device_id": "system", 00:07:27.742 "dma_device_type": 1 00:07:27.742 }, 00:07:27.742 { 00:07:27.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:27.742 "dma_device_type": 2 00:07:27.742 }, 00:07:27.742 { 00:07:27.742 "dma_device_id": "system", 00:07:27.742 "dma_device_type": 1 00:07:27.742 }, 00:07:27.742 { 00:07:27.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:27.742 "dma_device_type": 2 00:07:27.742 } 00:07:27.742 ], 00:07:27.742 "driver_specific": { 00:07:27.742 "raid": { 00:07:27.742 "uuid": "c326b647-184e-4706-a9c3-63b2fcb816c8", 00:07:27.742 "strip_size_kb": 64, 00:07:27.742 "state": "online", 00:07:27.742 "raid_level": "concat", 00:07:27.742 "superblock": true, 00:07:27.742 "num_base_bdevs": 2, 00:07:27.742 "num_base_bdevs_discovered": 2, 00:07:27.742 "num_base_bdevs_operational": 2, 00:07:27.742 "base_bdevs_list": [ 00:07:27.742 { 00:07:27.742 "name": "pt1", 00:07:27.742 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:27.742 "is_configured": true, 00:07:27.742 "data_offset": 2048, 00:07:27.742 "data_size": 63488 00:07:27.742 }, 00:07:27.742 { 00:07:27.742 "name": "pt2", 00:07:27.742 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:27.742 "is_configured": true, 00:07:27.742 "data_offset": 2048, 00:07:27.742 "data_size": 63488 00:07:27.742 } 00:07:27.742 ] 00:07:27.742 } 00:07:27.742 } 00:07:27.742 }' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:27.742 pt2' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.742 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:27.743 [2024-09-30 14:06:32.353062] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' c326b647-184e-4706-a9c3-63b2fcb816c8 '!=' c326b647-184e-4706-a9c3-63b2fcb816c8 ']' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74576 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 74576 ']' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 74576 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:27.743 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74576 00:07:28.002 killing process with pid 74576 00:07:28.002 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:28.002 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:28.002 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74576' 00:07:28.002 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 74576 00:07:28.002 [2024-09-30 14:06:32.426630] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:28.002 [2024-09-30 14:06:32.426709] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:28.002 [2024-09-30 14:06:32.426751] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:28.002 [2024-09-30 14:06:32.426761] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:28.002 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 74576 00:07:28.002 [2024-09-30 14:06:32.449199] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:28.261 14:06:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:28.261 00:07:28.261 real 0m3.359s 00:07:28.261 user 0m5.130s 00:07:28.261 sys 0m0.735s 00:07:28.261 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:28.261 14:06:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.261 ************************************ 00:07:28.261 END TEST raid_superblock_test 00:07:28.261 ************************************ 00:07:28.261 14:06:32 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:28.261 14:06:32 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:28.261 14:06:32 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:28.261 14:06:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:28.261 ************************************ 00:07:28.261 START TEST raid_read_error_test 00:07:28.261 ************************************ 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 read 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.L5wexud0KQ 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74772 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74772 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 74772 ']' 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:28.261 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:28.261 14:06:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.261 [2024-09-30 14:06:32.857214] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:28.261 [2024-09-30 14:06:32.857355] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74772 ] 00:07:28.520 [2024-09-30 14:06:32.988086] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:28.520 [2024-09-30 14:06:33.018678] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:28.520 [2024-09-30 14:06:33.064391] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:28.520 [2024-09-30 14:06:33.106634] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:28.520 [2024-09-30 14:06:33.106670] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.113 BaseBdev1_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.113 true 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.113 [2024-09-30 14:06:33.712467] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:29.113 [2024-09-30 14:06:33.712528] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:29.113 [2024-09-30 14:06:33.712545] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:29.113 [2024-09-30 14:06:33.712557] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:29.113 [2024-09-30 14:06:33.714579] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:29.113 [2024-09-30 14:06:33.714614] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:29.113 BaseBdev1 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.113 BaseBdev2_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.113 true 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.113 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.373 [2024-09-30 14:06:33.769905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:29.373 [2024-09-30 14:06:33.769981] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:29.373 [2024-09-30 14:06:33.770007] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:29.373 [2024-09-30 14:06:33.770026] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:29.373 [2024-09-30 14:06:33.772737] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:29.373 [2024-09-30 14:06:33.772784] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:29.373 BaseBdev2 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.373 [2024-09-30 14:06:33.781852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:29.373 [2024-09-30 14:06:33.783707] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:29.373 [2024-09-30 14:06:33.783880] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:29.373 [2024-09-30 14:06:33.783894] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:29.373 [2024-09-30 14:06:33.784134] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:29.373 [2024-09-30 14:06:33.784256] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:29.373 [2024-09-30 14:06:33.784265] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:29.373 [2024-09-30 14:06:33.784377] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.373 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.373 "name": "raid_bdev1", 00:07:29.373 "uuid": "c3e9fd59-e3b4-4493-a0c9-d74280170e2f", 00:07:29.373 "strip_size_kb": 64, 00:07:29.373 "state": "online", 00:07:29.373 "raid_level": "concat", 00:07:29.373 "superblock": true, 00:07:29.373 "num_base_bdevs": 2, 00:07:29.373 "num_base_bdevs_discovered": 2, 00:07:29.373 "num_base_bdevs_operational": 2, 00:07:29.373 "base_bdevs_list": [ 00:07:29.373 { 00:07:29.373 "name": "BaseBdev1", 00:07:29.373 "uuid": "d5755a4f-984b-5ae3-8395-4c38c4ccfcb5", 00:07:29.373 "is_configured": true, 00:07:29.373 "data_offset": 2048, 00:07:29.373 "data_size": 63488 00:07:29.373 }, 00:07:29.373 { 00:07:29.373 "name": "BaseBdev2", 00:07:29.373 "uuid": "6c14d09d-8bce-53f6-8998-6e3ab2f3e81c", 00:07:29.373 "is_configured": true, 00:07:29.373 "data_offset": 2048, 00:07:29.373 "data_size": 63488 00:07:29.373 } 00:07:29.373 ] 00:07:29.373 }' 00:07:29.374 14:06:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.374 14:06:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.632 14:06:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:29.632 14:06:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:29.891 [2024-09-30 14:06:34.289329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.829 "name": "raid_bdev1", 00:07:30.829 "uuid": "c3e9fd59-e3b4-4493-a0c9-d74280170e2f", 00:07:30.829 "strip_size_kb": 64, 00:07:30.829 "state": "online", 00:07:30.829 "raid_level": "concat", 00:07:30.829 "superblock": true, 00:07:30.829 "num_base_bdevs": 2, 00:07:30.829 "num_base_bdevs_discovered": 2, 00:07:30.829 "num_base_bdevs_operational": 2, 00:07:30.829 "base_bdevs_list": [ 00:07:30.829 { 00:07:30.829 "name": "BaseBdev1", 00:07:30.829 "uuid": "d5755a4f-984b-5ae3-8395-4c38c4ccfcb5", 00:07:30.829 "is_configured": true, 00:07:30.829 "data_offset": 2048, 00:07:30.829 "data_size": 63488 00:07:30.829 }, 00:07:30.829 { 00:07:30.829 "name": "BaseBdev2", 00:07:30.829 "uuid": "6c14d09d-8bce-53f6-8998-6e3ab2f3e81c", 00:07:30.829 "is_configured": true, 00:07:30.829 "data_offset": 2048, 00:07:30.829 "data_size": 63488 00:07:30.829 } 00:07:30.829 ] 00:07:30.829 }' 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.829 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.089 [2024-09-30 14:06:35.680906] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:31.089 [2024-09-30 14:06:35.680948] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:31.089 [2024-09-30 14:06:35.683367] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:31.089 [2024-09-30 14:06:35.683405] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:31.089 [2024-09-30 14:06:35.683434] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:31.089 [2024-09-30 14:06:35.683444] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:31.089 { 00:07:31.089 "results": [ 00:07:31.089 { 00:07:31.089 "job": "raid_bdev1", 00:07:31.089 "core_mask": "0x1", 00:07:31.089 "workload": "randrw", 00:07:31.089 "percentage": 50, 00:07:31.089 "status": "finished", 00:07:31.089 "queue_depth": 1, 00:07:31.089 "io_size": 131072, 00:07:31.089 "runtime": 1.392458, 00:07:31.089 "iops": 18056.558977003257, 00:07:31.089 "mibps": 2257.069872125407, 00:07:31.089 "io_failed": 1, 00:07:31.089 "io_timeout": 0, 00:07:31.089 "avg_latency_us": 76.54432126844571, 00:07:31.089 "min_latency_us": 24.370305676855896, 00:07:31.089 "max_latency_us": 1395.1441048034935 00:07:31.089 } 00:07:31.089 ], 00:07:31.089 "core_count": 1 00:07:31.089 } 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74772 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 74772 ']' 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 74772 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74772 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74772' 00:07:31.089 killing process with pid 74772 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 74772 00:07:31.089 [2024-09-30 14:06:35.730968] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:31.089 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 74772 00:07:31.348 [2024-09-30 14:06:35.746250] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.L5wexud0KQ 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:31.348 00:07:31.348 real 0m3.234s 00:07:31.348 user 0m4.071s 00:07:31.348 sys 0m0.519s 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:31.348 14:06:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.348 ************************************ 00:07:31.348 END TEST raid_read_error_test 00:07:31.348 ************************************ 00:07:31.607 14:06:36 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:07:31.607 14:06:36 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:31.607 14:06:36 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:31.607 14:06:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:31.607 ************************************ 00:07:31.607 START TEST raid_write_error_test 00:07:31.607 ************************************ 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 write 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.1OC72y2ftp 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74901 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74901 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 74901 ']' 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:31.607 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:31.607 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.607 [2024-09-30 14:06:36.178215] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:31.607 [2024-09-30 14:06:36.178506] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74901 ] 00:07:31.866 [2024-09-30 14:06:36.317303] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:31.866 [2024-09-30 14:06:36.344456] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.866 [2024-09-30 14:06:36.388794] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.866 [2024-09-30 14:06:36.431058] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:31.866 [2024-09-30 14:06:36.431099] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 BaseBdev1_malloc 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 true 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 [2024-09-30 14:06:37.013003] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:32.436 [2024-09-30 14:06:37.013055] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:32.436 [2024-09-30 14:06:37.013071] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:32.436 [2024-09-30 14:06:37.013082] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:32.436 [2024-09-30 14:06:37.015141] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:32.436 [2024-09-30 14:06:37.015248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:32.436 BaseBdev1 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 BaseBdev2_malloc 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 true 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 [2024-09-30 14:06:37.070452] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:32.436 [2024-09-30 14:06:37.070554] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:32.436 [2024-09-30 14:06:37.070581] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:32.436 [2024-09-30 14:06:37.070597] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:32.436 [2024-09-30 14:06:37.073843] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:32.436 [2024-09-30 14:06:37.073964] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:32.436 BaseBdev2 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.436 [2024-09-30 14:06:37.082637] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:32.436 [2024-09-30 14:06:37.084683] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:32.436 [2024-09-30 14:06:37.084913] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:32.436 [2024-09-30 14:06:37.084933] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:32.436 [2024-09-30 14:06:37.085218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:32.436 [2024-09-30 14:06:37.085368] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:32.436 [2024-09-30 14:06:37.085380] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:32.436 [2024-09-30 14:06:37.085522] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.436 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.696 "name": "raid_bdev1", 00:07:32.696 "uuid": "3afce00c-735e-474e-a2aa-2689b5fa2c91", 00:07:32.696 "strip_size_kb": 64, 00:07:32.696 "state": "online", 00:07:32.696 "raid_level": "concat", 00:07:32.696 "superblock": true, 00:07:32.696 "num_base_bdevs": 2, 00:07:32.696 "num_base_bdevs_discovered": 2, 00:07:32.696 "num_base_bdevs_operational": 2, 00:07:32.696 "base_bdevs_list": [ 00:07:32.696 { 00:07:32.696 "name": "BaseBdev1", 00:07:32.696 "uuid": "84cce4e7-b92a-5055-8f2f-dbfa08a56a39", 00:07:32.696 "is_configured": true, 00:07:32.696 "data_offset": 2048, 00:07:32.696 "data_size": 63488 00:07:32.696 }, 00:07:32.696 { 00:07:32.696 "name": "BaseBdev2", 00:07:32.696 "uuid": "dc84856f-7d1a-53e0-b50e-a4b25662d23f", 00:07:32.696 "is_configured": true, 00:07:32.696 "data_offset": 2048, 00:07:32.696 "data_size": 63488 00:07:32.696 } 00:07:32.696 ] 00:07:32.696 }' 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.696 14:06:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.955 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:32.955 14:06:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:32.955 [2024-09-30 14:06:37.606042] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:33.893 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:33.893 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.894 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.153 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.153 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.153 "name": "raid_bdev1", 00:07:34.153 "uuid": "3afce00c-735e-474e-a2aa-2689b5fa2c91", 00:07:34.153 "strip_size_kb": 64, 00:07:34.153 "state": "online", 00:07:34.153 "raid_level": "concat", 00:07:34.153 "superblock": true, 00:07:34.154 "num_base_bdevs": 2, 00:07:34.154 "num_base_bdevs_discovered": 2, 00:07:34.154 "num_base_bdevs_operational": 2, 00:07:34.154 "base_bdevs_list": [ 00:07:34.154 { 00:07:34.154 "name": "BaseBdev1", 00:07:34.154 "uuid": "84cce4e7-b92a-5055-8f2f-dbfa08a56a39", 00:07:34.154 "is_configured": true, 00:07:34.154 "data_offset": 2048, 00:07:34.154 "data_size": 63488 00:07:34.154 }, 00:07:34.154 { 00:07:34.154 "name": "BaseBdev2", 00:07:34.154 "uuid": "dc84856f-7d1a-53e0-b50e-a4b25662d23f", 00:07:34.154 "is_configured": true, 00:07:34.154 "data_offset": 2048, 00:07:34.154 "data_size": 63488 00:07:34.154 } 00:07:34.154 ] 00:07:34.154 }' 00:07:34.154 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.154 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.413 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:34.413 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.413 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.413 [2024-09-30 14:06:38.885134] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:34.413 [2024-09-30 14:06:38.885212] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:34.413 [2024-09-30 14:06:38.887653] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:34.413 [2024-09-30 14:06:38.887735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:34.413 [2024-09-30 14:06:38.887782] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:34.413 [2024-09-30 14:06:38.887824] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:34.413 { 00:07:34.413 "results": [ 00:07:34.413 { 00:07:34.413 "job": "raid_bdev1", 00:07:34.413 "core_mask": "0x1", 00:07:34.413 "workload": "randrw", 00:07:34.413 "percentage": 50, 00:07:34.413 "status": "finished", 00:07:34.414 "queue_depth": 1, 00:07:34.414 "io_size": 131072, 00:07:34.414 "runtime": 1.279876, 00:07:34.414 "iops": 18132.22530932684, 00:07:34.414 "mibps": 2266.528163665855, 00:07:34.414 "io_failed": 1, 00:07:34.414 "io_timeout": 0, 00:07:34.414 "avg_latency_us": 76.24100468292066, 00:07:34.414 "min_latency_us": 24.258515283842794, 00:07:34.414 "max_latency_us": 1359.3711790393013 00:07:34.414 } 00:07:34.414 ], 00:07:34.414 "core_count": 1 00:07:34.414 } 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74901 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 74901 ']' 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 74901 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74901 00:07:34.414 killing process with pid 74901 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74901' 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 74901 00:07:34.414 [2024-09-30 14:06:38.916992] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:34.414 14:06:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 74901 00:07:34.414 [2024-09-30 14:06:38.931842] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.1OC72y2ftp 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.78 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.78 != \0\.\0\0 ]] 00:07:34.672 00:07:34.672 real 0m3.109s 00:07:34.672 user 0m3.829s 00:07:34.672 sys 0m0.525s 00:07:34.672 ************************************ 00:07:34.672 END TEST raid_write_error_test 00:07:34.672 ************************************ 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:34.672 14:06:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.672 14:06:39 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:34.672 14:06:39 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:07:34.672 14:06:39 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:34.672 14:06:39 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:34.672 14:06:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:34.673 ************************************ 00:07:34.673 START TEST raid_state_function_test 00:07:34.673 ************************************ 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 false 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=75028 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75028' 00:07:34.673 Process raid pid: 75028 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 75028 00:07:34.673 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 75028 ']' 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:34.673 14:06:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.931 [2024-09-30 14:06:39.337473] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:34.931 [2024-09-30 14:06:39.337616] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:34.931 [2024-09-30 14:06:39.469446] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:34.931 [2024-09-30 14:06:39.498514] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:34.931 [2024-09-30 14:06:39.543862] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.931 [2024-09-30 14:06:39.585708] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.931 [2024-09-30 14:06:39.585736] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:35.499 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:35.499 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:35.499 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.499 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.499 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.758 [2024-09-30 14:06:40.158777] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:35.758 [2024-09-30 14:06:40.158825] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:35.758 [2024-09-30 14:06:40.158837] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.758 [2024-09-30 14:06:40.158844] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.758 "name": "Existed_Raid", 00:07:35.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.758 "strip_size_kb": 0, 00:07:35.758 "state": "configuring", 00:07:35.758 "raid_level": "raid1", 00:07:35.758 "superblock": false, 00:07:35.758 "num_base_bdevs": 2, 00:07:35.758 "num_base_bdevs_discovered": 0, 00:07:35.758 "num_base_bdevs_operational": 2, 00:07:35.758 "base_bdevs_list": [ 00:07:35.758 { 00:07:35.758 "name": "BaseBdev1", 00:07:35.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.758 "is_configured": false, 00:07:35.758 "data_offset": 0, 00:07:35.758 "data_size": 0 00:07:35.758 }, 00:07:35.758 { 00:07:35.758 "name": "BaseBdev2", 00:07:35.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.758 "is_configured": false, 00:07:35.758 "data_offset": 0, 00:07:35.758 "data_size": 0 00:07:35.758 } 00:07:35.758 ] 00:07:35.758 }' 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.758 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 [2024-09-30 14:06:40.514206] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:36.019 [2024-09-30 14:06:40.514277] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 [2024-09-30 14:06:40.522194] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:36.019 [2024-09-30 14:06:40.522265] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:36.019 [2024-09-30 14:06:40.522292] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:36.019 [2024-09-30 14:06:40.522311] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 [2024-09-30 14:06:40.538780] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:36.019 BaseBdev1 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.019 [ 00:07:36.019 { 00:07:36.019 "name": "BaseBdev1", 00:07:36.019 "aliases": [ 00:07:36.019 "9b559762-e872-4471-a454-88141bc4fa19" 00:07:36.019 ], 00:07:36.019 "product_name": "Malloc disk", 00:07:36.019 "block_size": 512, 00:07:36.019 "num_blocks": 65536, 00:07:36.019 "uuid": "9b559762-e872-4471-a454-88141bc4fa19", 00:07:36.019 "assigned_rate_limits": { 00:07:36.019 "rw_ios_per_sec": 0, 00:07:36.019 "rw_mbytes_per_sec": 0, 00:07:36.019 "r_mbytes_per_sec": 0, 00:07:36.019 "w_mbytes_per_sec": 0 00:07:36.019 }, 00:07:36.019 "claimed": true, 00:07:36.019 "claim_type": "exclusive_write", 00:07:36.019 "zoned": false, 00:07:36.019 "supported_io_types": { 00:07:36.019 "read": true, 00:07:36.019 "write": true, 00:07:36.019 "unmap": true, 00:07:36.019 "flush": true, 00:07:36.019 "reset": true, 00:07:36.019 "nvme_admin": false, 00:07:36.019 "nvme_io": false, 00:07:36.019 "nvme_io_md": false, 00:07:36.019 "write_zeroes": true, 00:07:36.019 "zcopy": true, 00:07:36.019 "get_zone_info": false, 00:07:36.019 "zone_management": false, 00:07:36.019 "zone_append": false, 00:07:36.019 "compare": false, 00:07:36.019 "compare_and_write": false, 00:07:36.019 "abort": true, 00:07:36.019 "seek_hole": false, 00:07:36.019 "seek_data": false, 00:07:36.019 "copy": true, 00:07:36.019 "nvme_iov_md": false 00:07:36.019 }, 00:07:36.019 "memory_domains": [ 00:07:36.019 { 00:07:36.019 "dma_device_id": "system", 00:07:36.019 "dma_device_type": 1 00:07:36.019 }, 00:07:36.019 { 00:07:36.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.019 "dma_device_type": 2 00:07:36.019 } 00:07:36.019 ], 00:07:36.019 "driver_specific": {} 00:07:36.019 } 00:07:36.019 ] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.019 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.020 "name": "Existed_Raid", 00:07:36.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.020 "strip_size_kb": 0, 00:07:36.020 "state": "configuring", 00:07:36.020 "raid_level": "raid1", 00:07:36.020 "superblock": false, 00:07:36.020 "num_base_bdevs": 2, 00:07:36.020 "num_base_bdevs_discovered": 1, 00:07:36.020 "num_base_bdevs_operational": 2, 00:07:36.020 "base_bdevs_list": [ 00:07:36.020 { 00:07:36.020 "name": "BaseBdev1", 00:07:36.020 "uuid": "9b559762-e872-4471-a454-88141bc4fa19", 00:07:36.020 "is_configured": true, 00:07:36.020 "data_offset": 0, 00:07:36.020 "data_size": 65536 00:07:36.020 }, 00:07:36.020 { 00:07:36.020 "name": "BaseBdev2", 00:07:36.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.020 "is_configured": false, 00:07:36.020 "data_offset": 0, 00:07:36.020 "data_size": 0 00:07:36.020 } 00:07:36.020 ] 00:07:36.020 }' 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.020 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.589 [2024-09-30 14:06:40.942123] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:36.589 [2024-09-30 14:06:40.942167] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.589 [2024-09-30 14:06:40.954116] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:36.589 [2024-09-30 14:06:40.955929] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:36.589 [2024-09-30 14:06:40.955969] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.589 "name": "Existed_Raid", 00:07:36.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.589 "strip_size_kb": 0, 00:07:36.589 "state": "configuring", 00:07:36.589 "raid_level": "raid1", 00:07:36.589 "superblock": false, 00:07:36.589 "num_base_bdevs": 2, 00:07:36.589 "num_base_bdevs_discovered": 1, 00:07:36.589 "num_base_bdevs_operational": 2, 00:07:36.589 "base_bdevs_list": [ 00:07:36.589 { 00:07:36.589 "name": "BaseBdev1", 00:07:36.589 "uuid": "9b559762-e872-4471-a454-88141bc4fa19", 00:07:36.589 "is_configured": true, 00:07:36.589 "data_offset": 0, 00:07:36.589 "data_size": 65536 00:07:36.589 }, 00:07:36.589 { 00:07:36.589 "name": "BaseBdev2", 00:07:36.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.589 "is_configured": false, 00:07:36.589 "data_offset": 0, 00:07:36.589 "data_size": 0 00:07:36.589 } 00:07:36.589 ] 00:07:36.589 }' 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.589 14:06:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.849 [2024-09-30 14:06:41.406250] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:36.849 [2024-09-30 14:06:41.406605] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:36.849 [2024-09-30 14:06:41.406804] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:36.849 [2024-09-30 14:06:41.407859] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:36.849 [2024-09-30 14:06:41.408530] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:36.849 [2024-09-30 14:06:41.408679] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:36.849 [2024-09-30 14:06:41.409578] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:36.849 BaseBdev2 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:36.849 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.850 [ 00:07:36.850 { 00:07:36.850 "name": "BaseBdev2", 00:07:36.850 "aliases": [ 00:07:36.850 "c55b0141-0466-409f-99dd-c43ee10426a8" 00:07:36.850 ], 00:07:36.850 "product_name": "Malloc disk", 00:07:36.850 "block_size": 512, 00:07:36.850 "num_blocks": 65536, 00:07:36.850 "uuid": "c55b0141-0466-409f-99dd-c43ee10426a8", 00:07:36.850 "assigned_rate_limits": { 00:07:36.850 "rw_ios_per_sec": 0, 00:07:36.850 "rw_mbytes_per_sec": 0, 00:07:36.850 "r_mbytes_per_sec": 0, 00:07:36.850 "w_mbytes_per_sec": 0 00:07:36.850 }, 00:07:36.850 "claimed": true, 00:07:36.850 "claim_type": "exclusive_write", 00:07:36.850 "zoned": false, 00:07:36.850 "supported_io_types": { 00:07:36.850 "read": true, 00:07:36.850 "write": true, 00:07:36.850 "unmap": true, 00:07:36.850 "flush": true, 00:07:36.850 "reset": true, 00:07:36.850 "nvme_admin": false, 00:07:36.850 "nvme_io": false, 00:07:36.850 "nvme_io_md": false, 00:07:36.850 "write_zeroes": true, 00:07:36.850 "zcopy": true, 00:07:36.850 "get_zone_info": false, 00:07:36.850 "zone_management": false, 00:07:36.850 "zone_append": false, 00:07:36.850 "compare": false, 00:07:36.850 "compare_and_write": false, 00:07:36.850 "abort": true, 00:07:36.850 "seek_hole": false, 00:07:36.850 "seek_data": false, 00:07:36.850 "copy": true, 00:07:36.850 "nvme_iov_md": false 00:07:36.850 }, 00:07:36.850 "memory_domains": [ 00:07:36.850 { 00:07:36.850 "dma_device_id": "system", 00:07:36.850 "dma_device_type": 1 00:07:36.850 }, 00:07:36.850 { 00:07:36.850 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.850 "dma_device_type": 2 00:07:36.850 } 00:07:36.850 ], 00:07:36.850 "driver_specific": {} 00:07:36.850 } 00:07:36.850 ] 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.850 "name": "Existed_Raid", 00:07:36.850 "uuid": "680d3be4-6cc4-43a8-b476-318c362030a4", 00:07:36.850 "strip_size_kb": 0, 00:07:36.850 "state": "online", 00:07:36.850 "raid_level": "raid1", 00:07:36.850 "superblock": false, 00:07:36.850 "num_base_bdevs": 2, 00:07:36.850 "num_base_bdevs_discovered": 2, 00:07:36.850 "num_base_bdevs_operational": 2, 00:07:36.850 "base_bdevs_list": [ 00:07:36.850 { 00:07:36.850 "name": "BaseBdev1", 00:07:36.850 "uuid": "9b559762-e872-4471-a454-88141bc4fa19", 00:07:36.850 "is_configured": true, 00:07:36.850 "data_offset": 0, 00:07:36.850 "data_size": 65536 00:07:36.850 }, 00:07:36.850 { 00:07:36.850 "name": "BaseBdev2", 00:07:36.850 "uuid": "c55b0141-0466-409f-99dd-c43ee10426a8", 00:07:36.850 "is_configured": true, 00:07:36.850 "data_offset": 0, 00:07:36.850 "data_size": 65536 00:07:36.850 } 00:07:36.850 ] 00:07:36.850 }' 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.850 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.419 [2024-09-30 14:06:41.909574] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:37.419 "name": "Existed_Raid", 00:07:37.419 "aliases": [ 00:07:37.419 "680d3be4-6cc4-43a8-b476-318c362030a4" 00:07:37.419 ], 00:07:37.419 "product_name": "Raid Volume", 00:07:37.419 "block_size": 512, 00:07:37.419 "num_blocks": 65536, 00:07:37.419 "uuid": "680d3be4-6cc4-43a8-b476-318c362030a4", 00:07:37.419 "assigned_rate_limits": { 00:07:37.419 "rw_ios_per_sec": 0, 00:07:37.419 "rw_mbytes_per_sec": 0, 00:07:37.419 "r_mbytes_per_sec": 0, 00:07:37.419 "w_mbytes_per_sec": 0 00:07:37.419 }, 00:07:37.419 "claimed": false, 00:07:37.419 "zoned": false, 00:07:37.419 "supported_io_types": { 00:07:37.419 "read": true, 00:07:37.419 "write": true, 00:07:37.419 "unmap": false, 00:07:37.419 "flush": false, 00:07:37.419 "reset": true, 00:07:37.419 "nvme_admin": false, 00:07:37.419 "nvme_io": false, 00:07:37.419 "nvme_io_md": false, 00:07:37.419 "write_zeroes": true, 00:07:37.419 "zcopy": false, 00:07:37.419 "get_zone_info": false, 00:07:37.419 "zone_management": false, 00:07:37.419 "zone_append": false, 00:07:37.419 "compare": false, 00:07:37.419 "compare_and_write": false, 00:07:37.419 "abort": false, 00:07:37.419 "seek_hole": false, 00:07:37.419 "seek_data": false, 00:07:37.419 "copy": false, 00:07:37.419 "nvme_iov_md": false 00:07:37.419 }, 00:07:37.419 "memory_domains": [ 00:07:37.419 { 00:07:37.419 "dma_device_id": "system", 00:07:37.419 "dma_device_type": 1 00:07:37.419 }, 00:07:37.419 { 00:07:37.419 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.419 "dma_device_type": 2 00:07:37.419 }, 00:07:37.419 { 00:07:37.419 "dma_device_id": "system", 00:07:37.419 "dma_device_type": 1 00:07:37.419 }, 00:07:37.419 { 00:07:37.419 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.419 "dma_device_type": 2 00:07:37.419 } 00:07:37.419 ], 00:07:37.419 "driver_specific": { 00:07:37.419 "raid": { 00:07:37.419 "uuid": "680d3be4-6cc4-43a8-b476-318c362030a4", 00:07:37.419 "strip_size_kb": 0, 00:07:37.419 "state": "online", 00:07:37.419 "raid_level": "raid1", 00:07:37.419 "superblock": false, 00:07:37.419 "num_base_bdevs": 2, 00:07:37.419 "num_base_bdevs_discovered": 2, 00:07:37.419 "num_base_bdevs_operational": 2, 00:07:37.419 "base_bdevs_list": [ 00:07:37.419 { 00:07:37.419 "name": "BaseBdev1", 00:07:37.419 "uuid": "9b559762-e872-4471-a454-88141bc4fa19", 00:07:37.419 "is_configured": true, 00:07:37.419 "data_offset": 0, 00:07:37.419 "data_size": 65536 00:07:37.419 }, 00:07:37.419 { 00:07:37.419 "name": "BaseBdev2", 00:07:37.419 "uuid": "c55b0141-0466-409f-99dd-c43ee10426a8", 00:07:37.419 "is_configured": true, 00:07:37.419 "data_offset": 0, 00:07:37.419 "data_size": 65536 00:07:37.419 } 00:07:37.419 ] 00:07:37.419 } 00:07:37.419 } 00:07:37.419 }' 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:37.419 BaseBdev2' 00:07:37.419 14:06:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.419 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.679 [2024-09-30 14:06:42.120997] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.679 "name": "Existed_Raid", 00:07:37.679 "uuid": "680d3be4-6cc4-43a8-b476-318c362030a4", 00:07:37.679 "strip_size_kb": 0, 00:07:37.679 "state": "online", 00:07:37.679 "raid_level": "raid1", 00:07:37.679 "superblock": false, 00:07:37.679 "num_base_bdevs": 2, 00:07:37.679 "num_base_bdevs_discovered": 1, 00:07:37.679 "num_base_bdevs_operational": 1, 00:07:37.679 "base_bdevs_list": [ 00:07:37.679 { 00:07:37.679 "name": null, 00:07:37.679 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:37.679 "is_configured": false, 00:07:37.679 "data_offset": 0, 00:07:37.679 "data_size": 65536 00:07:37.679 }, 00:07:37.679 { 00:07:37.679 "name": "BaseBdev2", 00:07:37.679 "uuid": "c55b0141-0466-409f-99dd-c43ee10426a8", 00:07:37.679 "is_configured": true, 00:07:37.679 "data_offset": 0, 00:07:37.679 "data_size": 65536 00:07:37.679 } 00:07:37.679 ] 00:07:37.679 }' 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.679 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:37.939 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.199 [2024-09-30 14:06:42.623230] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:38.199 [2024-09-30 14:06:42.623319] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:38.199 [2024-09-30 14:06:42.634567] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:38.199 [2024-09-30 14:06:42.634617] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:38.199 [2024-09-30 14:06:42.634629] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 75028 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 75028 ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 75028 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75028 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:38.199 killing process with pid 75028 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75028' 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 75028 00:07:38.199 [2024-09-30 14:06:42.726743] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:38.199 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 75028 00:07:38.199 [2024-09-30 14:06:42.727681] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:38.460 14:06:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:38.460 00:07:38.460 real 0m3.722s 00:07:38.460 user 0m5.833s 00:07:38.460 sys 0m0.723s 00:07:38.460 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:38.460 14:06:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.460 ************************************ 00:07:38.460 END TEST raid_state_function_test 00:07:38.460 ************************************ 00:07:38.460 14:06:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:38.460 14:06:43 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:38.460 14:06:43 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:38.460 14:06:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:38.460 ************************************ 00:07:38.460 START TEST raid_state_function_test_sb 00:07:38.460 ************************************ 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75265 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:38.460 Process raid pid: 75265 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75265' 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75265 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 75265 ']' 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:38.460 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:38.460 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.720 [2024-09-30 14:06:43.138064] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:38.720 [2024-09-30 14:06:43.138223] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:38.720 [2024-09-30 14:06:43.270298] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:38.720 [2024-09-30 14:06:43.299383] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:38.720 [2024-09-30 14:06:43.344462] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:38.980 [2024-09-30 14:06:43.385785] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:38.980 [2024-09-30 14:06:43.385823] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:39.548 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:39.548 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:39.548 14:06:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.548 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.548 14:06:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.548 [2024-09-30 14:06:44.002792] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:39.548 [2024-09-30 14:06:44.002838] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:39.548 [2024-09-30 14:06:44.002851] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.548 [2024-09-30 14:06:44.002858] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.548 "name": "Existed_Raid", 00:07:39.548 "uuid": "acb723ba-f251-41e7-ac52-a430aee2732b", 00:07:39.548 "strip_size_kb": 0, 00:07:39.548 "state": "configuring", 00:07:39.548 "raid_level": "raid1", 00:07:39.548 "superblock": true, 00:07:39.548 "num_base_bdevs": 2, 00:07:39.548 "num_base_bdevs_discovered": 0, 00:07:39.548 "num_base_bdevs_operational": 2, 00:07:39.548 "base_bdevs_list": [ 00:07:39.548 { 00:07:39.548 "name": "BaseBdev1", 00:07:39.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.548 "is_configured": false, 00:07:39.548 "data_offset": 0, 00:07:39.548 "data_size": 0 00:07:39.548 }, 00:07:39.548 { 00:07:39.548 "name": "BaseBdev2", 00:07:39.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.548 "is_configured": false, 00:07:39.548 "data_offset": 0, 00:07:39.548 "data_size": 0 00:07:39.548 } 00:07:39.548 ] 00:07:39.548 }' 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.548 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.807 [2024-09-30 14:06:44.430022] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:39.807 [2024-09-30 14:06:44.430094] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.807 [2024-09-30 14:06:44.442006] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:39.807 [2024-09-30 14:06:44.442072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:39.807 [2024-09-30 14:06:44.442099] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.807 [2024-09-30 14:06:44.442118] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.807 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.070 [2024-09-30 14:06:44.462893] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.070 BaseBdev1 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.070 [ 00:07:40.070 { 00:07:40.070 "name": "BaseBdev1", 00:07:40.070 "aliases": [ 00:07:40.070 "694ea325-0892-4050-9ce8-cd3a07bc05fa" 00:07:40.070 ], 00:07:40.070 "product_name": "Malloc disk", 00:07:40.070 "block_size": 512, 00:07:40.070 "num_blocks": 65536, 00:07:40.070 "uuid": "694ea325-0892-4050-9ce8-cd3a07bc05fa", 00:07:40.070 "assigned_rate_limits": { 00:07:40.070 "rw_ios_per_sec": 0, 00:07:40.070 "rw_mbytes_per_sec": 0, 00:07:40.070 "r_mbytes_per_sec": 0, 00:07:40.070 "w_mbytes_per_sec": 0 00:07:40.070 }, 00:07:40.070 "claimed": true, 00:07:40.070 "claim_type": "exclusive_write", 00:07:40.070 "zoned": false, 00:07:40.070 "supported_io_types": { 00:07:40.070 "read": true, 00:07:40.070 "write": true, 00:07:40.070 "unmap": true, 00:07:40.070 "flush": true, 00:07:40.070 "reset": true, 00:07:40.070 "nvme_admin": false, 00:07:40.070 "nvme_io": false, 00:07:40.070 "nvme_io_md": false, 00:07:40.070 "write_zeroes": true, 00:07:40.070 "zcopy": true, 00:07:40.070 "get_zone_info": false, 00:07:40.070 "zone_management": false, 00:07:40.070 "zone_append": false, 00:07:40.070 "compare": false, 00:07:40.070 "compare_and_write": false, 00:07:40.070 "abort": true, 00:07:40.070 "seek_hole": false, 00:07:40.070 "seek_data": false, 00:07:40.070 "copy": true, 00:07:40.070 "nvme_iov_md": false 00:07:40.070 }, 00:07:40.070 "memory_domains": [ 00:07:40.070 { 00:07:40.070 "dma_device_id": "system", 00:07:40.070 "dma_device_type": 1 00:07:40.070 }, 00:07:40.070 { 00:07:40.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.070 "dma_device_type": 2 00:07:40.070 } 00:07:40.070 ], 00:07:40.070 "driver_specific": {} 00:07:40.070 } 00:07:40.070 ] 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.070 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.071 "name": "Existed_Raid", 00:07:40.071 "uuid": "a8f5633e-785c-4f8c-9db4-42e376f1acae", 00:07:40.071 "strip_size_kb": 0, 00:07:40.071 "state": "configuring", 00:07:40.071 "raid_level": "raid1", 00:07:40.071 "superblock": true, 00:07:40.071 "num_base_bdevs": 2, 00:07:40.071 "num_base_bdevs_discovered": 1, 00:07:40.071 "num_base_bdevs_operational": 2, 00:07:40.071 "base_bdevs_list": [ 00:07:40.071 { 00:07:40.071 "name": "BaseBdev1", 00:07:40.071 "uuid": "694ea325-0892-4050-9ce8-cd3a07bc05fa", 00:07:40.071 "is_configured": true, 00:07:40.071 "data_offset": 2048, 00:07:40.071 "data_size": 63488 00:07:40.071 }, 00:07:40.071 { 00:07:40.071 "name": "BaseBdev2", 00:07:40.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.071 "is_configured": false, 00:07:40.071 "data_offset": 0, 00:07:40.071 "data_size": 0 00:07:40.071 } 00:07:40.071 ] 00:07:40.071 }' 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.071 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.330 [2024-09-30 14:06:44.906250] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:40.330 [2024-09-30 14:06:44.906297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.330 [2024-09-30 14:06:44.918236] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.330 [2024-09-30 14:06:44.920057] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:40.330 [2024-09-30 14:06:44.920143] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.330 "name": "Existed_Raid", 00:07:40.330 "uuid": "a62a970c-b7ff-4751-b3dd-9ca0c7b80161", 00:07:40.330 "strip_size_kb": 0, 00:07:40.330 "state": "configuring", 00:07:40.330 "raid_level": "raid1", 00:07:40.330 "superblock": true, 00:07:40.330 "num_base_bdevs": 2, 00:07:40.330 "num_base_bdevs_discovered": 1, 00:07:40.330 "num_base_bdevs_operational": 2, 00:07:40.330 "base_bdevs_list": [ 00:07:40.330 { 00:07:40.330 "name": "BaseBdev1", 00:07:40.330 "uuid": "694ea325-0892-4050-9ce8-cd3a07bc05fa", 00:07:40.330 "is_configured": true, 00:07:40.330 "data_offset": 2048, 00:07:40.330 "data_size": 63488 00:07:40.330 }, 00:07:40.330 { 00:07:40.330 "name": "BaseBdev2", 00:07:40.330 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.330 "is_configured": false, 00:07:40.330 "data_offset": 0, 00:07:40.330 "data_size": 0 00:07:40.330 } 00:07:40.330 ] 00:07:40.330 }' 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.330 14:06:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.900 [2024-09-30 14:06:45.368217] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:40.900 [2024-09-30 14:06:45.368599] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:40.900 [2024-09-30 14:06:45.368680] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:40.900 BaseBdev2 00:07:40.900 [2024-09-30 14:06:45.369172] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:40.900 [2024-09-30 14:06:45.369396] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:40.900 [2024-09-30 14:06:45.369500] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:40.900 [2024-09-30 14:06:45.369728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.900 [ 00:07:40.900 { 00:07:40.900 "name": "BaseBdev2", 00:07:40.900 "aliases": [ 00:07:40.900 "032beece-e693-4b0b-a420-2058d6af4a6b" 00:07:40.900 ], 00:07:40.900 "product_name": "Malloc disk", 00:07:40.900 "block_size": 512, 00:07:40.900 "num_blocks": 65536, 00:07:40.900 "uuid": "032beece-e693-4b0b-a420-2058d6af4a6b", 00:07:40.900 "assigned_rate_limits": { 00:07:40.900 "rw_ios_per_sec": 0, 00:07:40.900 "rw_mbytes_per_sec": 0, 00:07:40.900 "r_mbytes_per_sec": 0, 00:07:40.900 "w_mbytes_per_sec": 0 00:07:40.900 }, 00:07:40.900 "claimed": true, 00:07:40.900 "claim_type": "exclusive_write", 00:07:40.900 "zoned": false, 00:07:40.900 "supported_io_types": { 00:07:40.900 "read": true, 00:07:40.900 "write": true, 00:07:40.900 "unmap": true, 00:07:40.900 "flush": true, 00:07:40.900 "reset": true, 00:07:40.900 "nvme_admin": false, 00:07:40.900 "nvme_io": false, 00:07:40.900 "nvme_io_md": false, 00:07:40.900 "write_zeroes": true, 00:07:40.900 "zcopy": true, 00:07:40.900 "get_zone_info": false, 00:07:40.900 "zone_management": false, 00:07:40.900 "zone_append": false, 00:07:40.900 "compare": false, 00:07:40.900 "compare_and_write": false, 00:07:40.900 "abort": true, 00:07:40.900 "seek_hole": false, 00:07:40.900 "seek_data": false, 00:07:40.900 "copy": true, 00:07:40.900 "nvme_iov_md": false 00:07:40.900 }, 00:07:40.900 "memory_domains": [ 00:07:40.900 { 00:07:40.900 "dma_device_id": "system", 00:07:40.900 "dma_device_type": 1 00:07:40.900 }, 00:07:40.900 { 00:07:40.900 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.900 "dma_device_type": 2 00:07:40.900 } 00:07:40.900 ], 00:07:40.900 "driver_specific": {} 00:07:40.900 } 00:07:40.900 ] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.900 "name": "Existed_Raid", 00:07:40.900 "uuid": "a62a970c-b7ff-4751-b3dd-9ca0c7b80161", 00:07:40.900 "strip_size_kb": 0, 00:07:40.900 "state": "online", 00:07:40.900 "raid_level": "raid1", 00:07:40.900 "superblock": true, 00:07:40.900 "num_base_bdevs": 2, 00:07:40.900 "num_base_bdevs_discovered": 2, 00:07:40.900 "num_base_bdevs_operational": 2, 00:07:40.900 "base_bdevs_list": [ 00:07:40.900 { 00:07:40.900 "name": "BaseBdev1", 00:07:40.900 "uuid": "694ea325-0892-4050-9ce8-cd3a07bc05fa", 00:07:40.900 "is_configured": true, 00:07:40.900 "data_offset": 2048, 00:07:40.900 "data_size": 63488 00:07:40.900 }, 00:07:40.900 { 00:07:40.900 "name": "BaseBdev2", 00:07:40.900 "uuid": "032beece-e693-4b0b-a420-2058d6af4a6b", 00:07:40.900 "is_configured": true, 00:07:40.900 "data_offset": 2048, 00:07:40.900 "data_size": 63488 00:07:40.900 } 00:07:40.900 ] 00:07:40.900 }' 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.900 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.471 [2024-09-30 14:06:45.847665] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:41.471 "name": "Existed_Raid", 00:07:41.471 "aliases": [ 00:07:41.471 "a62a970c-b7ff-4751-b3dd-9ca0c7b80161" 00:07:41.471 ], 00:07:41.471 "product_name": "Raid Volume", 00:07:41.471 "block_size": 512, 00:07:41.471 "num_blocks": 63488, 00:07:41.471 "uuid": "a62a970c-b7ff-4751-b3dd-9ca0c7b80161", 00:07:41.471 "assigned_rate_limits": { 00:07:41.471 "rw_ios_per_sec": 0, 00:07:41.471 "rw_mbytes_per_sec": 0, 00:07:41.471 "r_mbytes_per_sec": 0, 00:07:41.471 "w_mbytes_per_sec": 0 00:07:41.471 }, 00:07:41.471 "claimed": false, 00:07:41.471 "zoned": false, 00:07:41.471 "supported_io_types": { 00:07:41.471 "read": true, 00:07:41.471 "write": true, 00:07:41.471 "unmap": false, 00:07:41.471 "flush": false, 00:07:41.471 "reset": true, 00:07:41.471 "nvme_admin": false, 00:07:41.471 "nvme_io": false, 00:07:41.471 "nvme_io_md": false, 00:07:41.471 "write_zeroes": true, 00:07:41.471 "zcopy": false, 00:07:41.471 "get_zone_info": false, 00:07:41.471 "zone_management": false, 00:07:41.471 "zone_append": false, 00:07:41.471 "compare": false, 00:07:41.471 "compare_and_write": false, 00:07:41.471 "abort": false, 00:07:41.471 "seek_hole": false, 00:07:41.471 "seek_data": false, 00:07:41.471 "copy": false, 00:07:41.471 "nvme_iov_md": false 00:07:41.471 }, 00:07:41.471 "memory_domains": [ 00:07:41.471 { 00:07:41.471 "dma_device_id": "system", 00:07:41.471 "dma_device_type": 1 00:07:41.471 }, 00:07:41.471 { 00:07:41.471 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.471 "dma_device_type": 2 00:07:41.471 }, 00:07:41.471 { 00:07:41.471 "dma_device_id": "system", 00:07:41.471 "dma_device_type": 1 00:07:41.471 }, 00:07:41.471 { 00:07:41.471 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.471 "dma_device_type": 2 00:07:41.471 } 00:07:41.471 ], 00:07:41.471 "driver_specific": { 00:07:41.471 "raid": { 00:07:41.471 "uuid": "a62a970c-b7ff-4751-b3dd-9ca0c7b80161", 00:07:41.471 "strip_size_kb": 0, 00:07:41.471 "state": "online", 00:07:41.471 "raid_level": "raid1", 00:07:41.471 "superblock": true, 00:07:41.471 "num_base_bdevs": 2, 00:07:41.471 "num_base_bdevs_discovered": 2, 00:07:41.471 "num_base_bdevs_operational": 2, 00:07:41.471 "base_bdevs_list": [ 00:07:41.471 { 00:07:41.471 "name": "BaseBdev1", 00:07:41.471 "uuid": "694ea325-0892-4050-9ce8-cd3a07bc05fa", 00:07:41.471 "is_configured": true, 00:07:41.471 "data_offset": 2048, 00:07:41.471 "data_size": 63488 00:07:41.471 }, 00:07:41.471 { 00:07:41.471 "name": "BaseBdev2", 00:07:41.471 "uuid": "032beece-e693-4b0b-a420-2058d6af4a6b", 00:07:41.471 "is_configured": true, 00:07:41.471 "data_offset": 2048, 00:07:41.471 "data_size": 63488 00:07:41.471 } 00:07:41.471 ] 00:07:41.471 } 00:07:41.471 } 00:07:41.471 }' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:41.471 BaseBdev2' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.471 14:06:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.471 [2024-09-30 14:06:46.087036] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:41.471 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.472 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.731 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:41.731 "name": "Existed_Raid", 00:07:41.731 "uuid": "a62a970c-b7ff-4751-b3dd-9ca0c7b80161", 00:07:41.731 "strip_size_kb": 0, 00:07:41.731 "state": "online", 00:07:41.731 "raid_level": "raid1", 00:07:41.731 "superblock": true, 00:07:41.731 "num_base_bdevs": 2, 00:07:41.731 "num_base_bdevs_discovered": 1, 00:07:41.731 "num_base_bdevs_operational": 1, 00:07:41.731 "base_bdevs_list": [ 00:07:41.731 { 00:07:41.731 "name": null, 00:07:41.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:41.731 "is_configured": false, 00:07:41.731 "data_offset": 0, 00:07:41.731 "data_size": 63488 00:07:41.731 }, 00:07:41.731 { 00:07:41.731 "name": "BaseBdev2", 00:07:41.731 "uuid": "032beece-e693-4b0b-a420-2058d6af4a6b", 00:07:41.731 "is_configured": true, 00:07:41.731 "data_offset": 2048, 00:07:41.731 "data_size": 63488 00:07:41.731 } 00:07:41.731 ] 00:07:41.731 }' 00:07:41.731 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:41.731 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.991 [2024-09-30 14:06:46.593380] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:41.991 [2024-09-30 14:06:46.593535] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:41.991 [2024-09-30 14:06:46.604900] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:41.991 [2024-09-30 14:06:46.605021] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:41.991 [2024-09-30 14:06:46.605063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.991 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75265 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 75265 ']' 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 75265 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75265 00:07:42.251 killing process with pid 75265 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:42.251 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75265' 00:07:42.252 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 75265 00:07:42.252 [2024-09-30 14:06:46.699669] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:42.252 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 75265 00:07:42.252 [2024-09-30 14:06:46.700629] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:42.512 14:06:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:42.512 00:07:42.512 real 0m3.898s 00:07:42.512 user 0m6.142s 00:07:42.512 sys 0m0.740s 00:07:42.512 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:42.512 14:06:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.512 ************************************ 00:07:42.512 END TEST raid_state_function_test_sb 00:07:42.512 ************************************ 00:07:42.512 14:06:47 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:42.512 14:06:47 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:42.512 14:06:47 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:42.512 14:06:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:42.512 ************************************ 00:07:42.512 START TEST raid_superblock_test 00:07:42.512 ************************************ 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75506 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75506 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 75506 ']' 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:42.512 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:42.512 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.512 [2024-09-30 14:06:47.098046] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:42.512 [2024-09-30 14:06:47.098243] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75506 ] 00:07:42.772 [2024-09-30 14:06:47.228253] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:42.772 [2024-09-30 14:06:47.252755] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:42.772 [2024-09-30 14:06:47.296579] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:42.772 [2024-09-30 14:06:47.337616] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:42.772 [2024-09-30 14:06:47.337649] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.341 malloc1 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.341 [2024-09-30 14:06:47.942999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:43.341 [2024-09-30 14:06:47.943105] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.341 [2024-09-30 14:06:47.943148] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:43.341 [2024-09-30 14:06:47.943178] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.341 [2024-09-30 14:06:47.945228] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.341 [2024-09-30 14:06:47.945295] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:43.341 pt1 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.341 malloc2 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.341 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.341 [2024-09-30 14:06:47.991803] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:43.341 [2024-09-30 14:06:47.991945] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.341 [2024-09-30 14:06:47.992002] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:43.341 [2024-09-30 14:06:47.992032] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.601 [2024-09-30 14:06:47.996878] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.601 [2024-09-30 14:06:47.996944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:43.601 pt2 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.601 14:06:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.601 [2024-09-30 14:06:48.005199] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:43.601 [2024-09-30 14:06:48.008031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:43.601 [2024-09-30 14:06:48.008291] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:07:43.601 [2024-09-30 14:06:48.008357] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:43.601 [2024-09-30 14:06:48.008800] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:43.601 [2024-09-30 14:06:48.009039] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:07:43.601 [2024-09-30 14:06:48.009105] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:07:43.601 [2024-09-30 14:06:48.009377] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.601 "name": "raid_bdev1", 00:07:43.601 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:43.601 "strip_size_kb": 0, 00:07:43.601 "state": "online", 00:07:43.601 "raid_level": "raid1", 00:07:43.601 "superblock": true, 00:07:43.601 "num_base_bdevs": 2, 00:07:43.601 "num_base_bdevs_discovered": 2, 00:07:43.601 "num_base_bdevs_operational": 2, 00:07:43.601 "base_bdevs_list": [ 00:07:43.601 { 00:07:43.601 "name": "pt1", 00:07:43.601 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.601 "is_configured": true, 00:07:43.601 "data_offset": 2048, 00:07:43.601 "data_size": 63488 00:07:43.601 }, 00:07:43.601 { 00:07:43.601 "name": "pt2", 00:07:43.601 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.601 "is_configured": true, 00:07:43.601 "data_offset": 2048, 00:07:43.601 "data_size": 63488 00:07:43.601 } 00:07:43.601 ] 00:07:43.601 }' 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.601 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.857 [2024-09-30 14:06:48.416897] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.857 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:43.857 "name": "raid_bdev1", 00:07:43.857 "aliases": [ 00:07:43.857 "a3388d92-0791-42ad-aaa7-66b388e8ccfa" 00:07:43.857 ], 00:07:43.857 "product_name": "Raid Volume", 00:07:43.857 "block_size": 512, 00:07:43.857 "num_blocks": 63488, 00:07:43.857 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:43.857 "assigned_rate_limits": { 00:07:43.857 "rw_ios_per_sec": 0, 00:07:43.857 "rw_mbytes_per_sec": 0, 00:07:43.858 "r_mbytes_per_sec": 0, 00:07:43.858 "w_mbytes_per_sec": 0 00:07:43.858 }, 00:07:43.858 "claimed": false, 00:07:43.858 "zoned": false, 00:07:43.858 "supported_io_types": { 00:07:43.858 "read": true, 00:07:43.858 "write": true, 00:07:43.858 "unmap": false, 00:07:43.858 "flush": false, 00:07:43.858 "reset": true, 00:07:43.858 "nvme_admin": false, 00:07:43.858 "nvme_io": false, 00:07:43.858 "nvme_io_md": false, 00:07:43.858 "write_zeroes": true, 00:07:43.858 "zcopy": false, 00:07:43.858 "get_zone_info": false, 00:07:43.858 "zone_management": false, 00:07:43.858 "zone_append": false, 00:07:43.858 "compare": false, 00:07:43.858 "compare_and_write": false, 00:07:43.858 "abort": false, 00:07:43.858 "seek_hole": false, 00:07:43.858 "seek_data": false, 00:07:43.858 "copy": false, 00:07:43.858 "nvme_iov_md": false 00:07:43.858 }, 00:07:43.858 "memory_domains": [ 00:07:43.858 { 00:07:43.858 "dma_device_id": "system", 00:07:43.858 "dma_device_type": 1 00:07:43.858 }, 00:07:43.858 { 00:07:43.858 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.858 "dma_device_type": 2 00:07:43.858 }, 00:07:43.858 { 00:07:43.858 "dma_device_id": "system", 00:07:43.858 "dma_device_type": 1 00:07:43.858 }, 00:07:43.858 { 00:07:43.858 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.858 "dma_device_type": 2 00:07:43.858 } 00:07:43.858 ], 00:07:43.858 "driver_specific": { 00:07:43.858 "raid": { 00:07:43.858 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:43.858 "strip_size_kb": 0, 00:07:43.858 "state": "online", 00:07:43.858 "raid_level": "raid1", 00:07:43.858 "superblock": true, 00:07:43.858 "num_base_bdevs": 2, 00:07:43.858 "num_base_bdevs_discovered": 2, 00:07:43.858 "num_base_bdevs_operational": 2, 00:07:43.858 "base_bdevs_list": [ 00:07:43.858 { 00:07:43.858 "name": "pt1", 00:07:43.858 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.858 "is_configured": true, 00:07:43.858 "data_offset": 2048, 00:07:43.858 "data_size": 63488 00:07:43.858 }, 00:07:43.858 { 00:07:43.858 "name": "pt2", 00:07:43.858 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.858 "is_configured": true, 00:07:43.858 "data_offset": 2048, 00:07:43.858 "data_size": 63488 00:07:43.858 } 00:07:43.858 ] 00:07:43.858 } 00:07:43.858 } 00:07:43.858 }' 00:07:43.858 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:43.858 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:43.858 pt2' 00:07:43.858 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:44.116 [2024-09-30 14:06:48.652416] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a3388d92-0791-42ad-aaa7-66b388e8ccfa 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a3388d92-0791-42ad-aaa7-66b388e8ccfa ']' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 [2024-09-30 14:06:48.680125] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.116 [2024-09-30 14:06:48.680184] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:44.116 [2024-09-30 14:06:48.680256] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:44.116 [2024-09-30 14:06:48.680323] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:44.116 [2024-09-30 14:06:48.680336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:44.116 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.389 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.390 [2024-09-30 14:06:48.807930] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:44.390 [2024-09-30 14:06:48.809748] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:44.390 [2024-09-30 14:06:48.809808] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:44.390 [2024-09-30 14:06:48.809850] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:44.390 [2024-09-30 14:06:48.809864] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.390 [2024-09-30 14:06:48.809873] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:07:44.390 request: 00:07:44.390 { 00:07:44.390 "name": "raid_bdev1", 00:07:44.390 "raid_level": "raid1", 00:07:44.390 "base_bdevs": [ 00:07:44.390 "malloc1", 00:07:44.390 "malloc2" 00:07:44.390 ], 00:07:44.390 "superblock": false, 00:07:44.390 "method": "bdev_raid_create", 00:07:44.390 "req_id": 1 00:07:44.390 } 00:07:44.390 Got JSON-RPC error response 00:07:44.390 response: 00:07:44.390 { 00:07:44.390 "code": -17, 00:07:44.390 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:44.390 } 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.390 [2024-09-30 14:06:48.871809] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:44.390 [2024-09-30 14:06:48.871893] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.390 [2024-09-30 14:06:48.871924] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:44.390 [2024-09-30 14:06:48.871954] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.390 [2024-09-30 14:06:48.873945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.390 [2024-09-30 14:06:48.874015] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:44.390 [2024-09-30 14:06:48.874114] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:44.390 [2024-09-30 14:06:48.874178] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:44.390 pt1 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.390 "name": "raid_bdev1", 00:07:44.390 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:44.390 "strip_size_kb": 0, 00:07:44.390 "state": "configuring", 00:07:44.390 "raid_level": "raid1", 00:07:44.390 "superblock": true, 00:07:44.390 "num_base_bdevs": 2, 00:07:44.390 "num_base_bdevs_discovered": 1, 00:07:44.390 "num_base_bdevs_operational": 2, 00:07:44.390 "base_bdevs_list": [ 00:07:44.390 { 00:07:44.390 "name": "pt1", 00:07:44.390 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.390 "is_configured": true, 00:07:44.390 "data_offset": 2048, 00:07:44.390 "data_size": 63488 00:07:44.390 }, 00:07:44.390 { 00:07:44.390 "name": null, 00:07:44.390 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.390 "is_configured": false, 00:07:44.390 "data_offset": 2048, 00:07:44.390 "data_size": 63488 00:07:44.390 } 00:07:44.390 ] 00:07:44.390 }' 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.390 14:06:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.650 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.650 [2024-09-30 14:06:49.303070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:44.650 [2024-09-30 14:06:49.303179] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.650 [2024-09-30 14:06:49.303203] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:44.650 [2024-09-30 14:06:49.303213] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.650 [2024-09-30 14:06:49.303604] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.650 [2024-09-30 14:06:49.303629] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:44.650 [2024-09-30 14:06:49.303696] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:44.650 [2024-09-30 14:06:49.303719] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:44.650 [2024-09-30 14:06:49.303813] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:44.650 [2024-09-30 14:06:49.303825] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:44.650 [2024-09-30 14:06:49.304036] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:44.650 [2024-09-30 14:06:49.304157] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:44.650 [2024-09-30 14:06:49.304166] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:44.650 [2024-09-30 14:06:49.304263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.908 pt2 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.908 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.908 "name": "raid_bdev1", 00:07:44.908 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:44.908 "strip_size_kb": 0, 00:07:44.908 "state": "online", 00:07:44.908 "raid_level": "raid1", 00:07:44.908 "superblock": true, 00:07:44.908 "num_base_bdevs": 2, 00:07:44.908 "num_base_bdevs_discovered": 2, 00:07:44.908 "num_base_bdevs_operational": 2, 00:07:44.908 "base_bdevs_list": [ 00:07:44.908 { 00:07:44.908 "name": "pt1", 00:07:44.909 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.909 "is_configured": true, 00:07:44.909 "data_offset": 2048, 00:07:44.909 "data_size": 63488 00:07:44.909 }, 00:07:44.909 { 00:07:44.909 "name": "pt2", 00:07:44.909 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.909 "is_configured": true, 00:07:44.909 "data_offset": 2048, 00:07:44.909 "data_size": 63488 00:07:44.909 } 00:07:44.909 ] 00:07:44.909 }' 00:07:44.909 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.909 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.168 [2024-09-30 14:06:49.702641] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:45.168 "name": "raid_bdev1", 00:07:45.168 "aliases": [ 00:07:45.168 "a3388d92-0791-42ad-aaa7-66b388e8ccfa" 00:07:45.168 ], 00:07:45.168 "product_name": "Raid Volume", 00:07:45.168 "block_size": 512, 00:07:45.168 "num_blocks": 63488, 00:07:45.168 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:45.168 "assigned_rate_limits": { 00:07:45.168 "rw_ios_per_sec": 0, 00:07:45.168 "rw_mbytes_per_sec": 0, 00:07:45.168 "r_mbytes_per_sec": 0, 00:07:45.168 "w_mbytes_per_sec": 0 00:07:45.168 }, 00:07:45.168 "claimed": false, 00:07:45.168 "zoned": false, 00:07:45.168 "supported_io_types": { 00:07:45.168 "read": true, 00:07:45.168 "write": true, 00:07:45.168 "unmap": false, 00:07:45.168 "flush": false, 00:07:45.168 "reset": true, 00:07:45.168 "nvme_admin": false, 00:07:45.168 "nvme_io": false, 00:07:45.168 "nvme_io_md": false, 00:07:45.168 "write_zeroes": true, 00:07:45.168 "zcopy": false, 00:07:45.168 "get_zone_info": false, 00:07:45.168 "zone_management": false, 00:07:45.168 "zone_append": false, 00:07:45.168 "compare": false, 00:07:45.168 "compare_and_write": false, 00:07:45.168 "abort": false, 00:07:45.168 "seek_hole": false, 00:07:45.168 "seek_data": false, 00:07:45.168 "copy": false, 00:07:45.168 "nvme_iov_md": false 00:07:45.168 }, 00:07:45.168 "memory_domains": [ 00:07:45.168 { 00:07:45.168 "dma_device_id": "system", 00:07:45.168 "dma_device_type": 1 00:07:45.168 }, 00:07:45.168 { 00:07:45.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.168 "dma_device_type": 2 00:07:45.168 }, 00:07:45.168 { 00:07:45.168 "dma_device_id": "system", 00:07:45.168 "dma_device_type": 1 00:07:45.168 }, 00:07:45.168 { 00:07:45.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.168 "dma_device_type": 2 00:07:45.168 } 00:07:45.168 ], 00:07:45.168 "driver_specific": { 00:07:45.168 "raid": { 00:07:45.168 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:45.168 "strip_size_kb": 0, 00:07:45.168 "state": "online", 00:07:45.168 "raid_level": "raid1", 00:07:45.168 "superblock": true, 00:07:45.168 "num_base_bdevs": 2, 00:07:45.168 "num_base_bdevs_discovered": 2, 00:07:45.168 "num_base_bdevs_operational": 2, 00:07:45.168 "base_bdevs_list": [ 00:07:45.168 { 00:07:45.168 "name": "pt1", 00:07:45.168 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:45.168 "is_configured": true, 00:07:45.168 "data_offset": 2048, 00:07:45.168 "data_size": 63488 00:07:45.168 }, 00:07:45.168 { 00:07:45.168 "name": "pt2", 00:07:45.168 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.168 "is_configured": true, 00:07:45.168 "data_offset": 2048, 00:07:45.168 "data_size": 63488 00:07:45.168 } 00:07:45.168 ] 00:07:45.168 } 00:07:45.168 } 00:07:45.168 }' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:45.168 pt2' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.168 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.427 [2024-09-30 14:06:49.906255] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a3388d92-0791-42ad-aaa7-66b388e8ccfa '!=' a3388d92-0791-42ad-aaa7-66b388e8ccfa ']' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.427 [2024-09-30 14:06:49.953987] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.427 14:06:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.427 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.427 "name": "raid_bdev1", 00:07:45.427 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:45.427 "strip_size_kb": 0, 00:07:45.427 "state": "online", 00:07:45.427 "raid_level": "raid1", 00:07:45.427 "superblock": true, 00:07:45.427 "num_base_bdevs": 2, 00:07:45.427 "num_base_bdevs_discovered": 1, 00:07:45.427 "num_base_bdevs_operational": 1, 00:07:45.427 "base_bdevs_list": [ 00:07:45.427 { 00:07:45.427 "name": null, 00:07:45.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:45.427 "is_configured": false, 00:07:45.427 "data_offset": 0, 00:07:45.427 "data_size": 63488 00:07:45.427 }, 00:07:45.427 { 00:07:45.427 "name": "pt2", 00:07:45.427 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.427 "is_configured": true, 00:07:45.427 "data_offset": 2048, 00:07:45.427 "data_size": 63488 00:07:45.427 } 00:07:45.427 ] 00:07:45.427 }' 00:07:45.427 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.427 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.994 [2024-09-30 14:06:50.385233] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:45.994 [2024-09-30 14:06:50.385257] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:45.994 [2024-09-30 14:06:50.385324] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.994 [2024-09-30 14:06:50.385366] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.994 [2024-09-30 14:06:50.385376] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.994 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.995 [2024-09-30 14:06:50.453109] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:45.995 [2024-09-30 14:06:50.453159] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.995 [2024-09-30 14:06:50.453175] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:07:45.995 [2024-09-30 14:06:50.453185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.995 [2024-09-30 14:06:50.455280] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.995 [2024-09-30 14:06:50.455318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:45.995 [2024-09-30 14:06:50.455389] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:45.995 [2024-09-30 14:06:50.455423] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:45.995 [2024-09-30 14:06:50.455504] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:07:45.995 [2024-09-30 14:06:50.455518] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:45.995 [2024-09-30 14:06:50.455724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:45.995 [2024-09-30 14:06:50.455835] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:07:45.995 [2024-09-30 14:06:50.455843] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:07:45.995 [2024-09-30 14:06:50.455940] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:45.995 pt2 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.995 "name": "raid_bdev1", 00:07:45.995 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:45.995 "strip_size_kb": 0, 00:07:45.995 "state": "online", 00:07:45.995 "raid_level": "raid1", 00:07:45.995 "superblock": true, 00:07:45.995 "num_base_bdevs": 2, 00:07:45.995 "num_base_bdevs_discovered": 1, 00:07:45.995 "num_base_bdevs_operational": 1, 00:07:45.995 "base_bdevs_list": [ 00:07:45.995 { 00:07:45.995 "name": null, 00:07:45.995 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:45.995 "is_configured": false, 00:07:45.995 "data_offset": 2048, 00:07:45.995 "data_size": 63488 00:07:45.995 }, 00:07:45.995 { 00:07:45.995 "name": "pt2", 00:07:45.995 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.995 "is_configured": true, 00:07:45.995 "data_offset": 2048, 00:07:45.995 "data_size": 63488 00:07:45.995 } 00:07:45.995 ] 00:07:45.995 }' 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.995 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.255 [2024-09-30 14:06:50.852446] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:46.255 [2024-09-30 14:06:50.852476] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:46.255 [2024-09-30 14:06:50.852561] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:46.255 [2024-09-30 14:06:50.852609] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:46.255 [2024-09-30 14:06:50.852617] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.255 [2024-09-30 14:06:50.900344] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:46.255 [2024-09-30 14:06:50.900438] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.255 [2024-09-30 14:06:50.900468] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:07:46.255 [2024-09-30 14:06:50.900489] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.255 [2024-09-30 14:06:50.902618] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.255 [2024-09-30 14:06:50.902654] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:46.255 [2024-09-30 14:06:50.902728] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:46.255 [2024-09-30 14:06:50.902759] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:46.255 [2024-09-30 14:06:50.902855] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:07:46.255 [2024-09-30 14:06:50.902865] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:46.255 [2024-09-30 14:06:50.902883] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:07:46.255 [2024-09-30 14:06:50.902926] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:46.255 [2024-09-30 14:06:50.902999] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:07:46.255 [2024-09-30 14:06:50.903007] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:46.255 [2024-09-30 14:06:50.903227] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:07:46.255 [2024-09-30 14:06:50.903345] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:07:46.255 [2024-09-30 14:06:50.903357] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:07:46.255 [2024-09-30 14:06:50.903463] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:46.255 pt1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:46.255 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:46.514 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:46.515 "name": "raid_bdev1", 00:07:46.515 "uuid": "a3388d92-0791-42ad-aaa7-66b388e8ccfa", 00:07:46.515 "strip_size_kb": 0, 00:07:46.515 "state": "online", 00:07:46.515 "raid_level": "raid1", 00:07:46.515 "superblock": true, 00:07:46.515 "num_base_bdevs": 2, 00:07:46.515 "num_base_bdevs_discovered": 1, 00:07:46.515 "num_base_bdevs_operational": 1, 00:07:46.515 "base_bdevs_list": [ 00:07:46.515 { 00:07:46.515 "name": null, 00:07:46.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:46.515 "is_configured": false, 00:07:46.515 "data_offset": 2048, 00:07:46.515 "data_size": 63488 00:07:46.515 }, 00:07:46.515 { 00:07:46.515 "name": "pt2", 00:07:46.515 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:46.515 "is_configured": true, 00:07:46.515 "data_offset": 2048, 00:07:46.515 "data_size": 63488 00:07:46.515 } 00:07:46.515 ] 00:07:46.515 }' 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:46.515 14:06:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.775 [2024-09-30 14:06:51.327844] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' a3388d92-0791-42ad-aaa7-66b388e8ccfa '!=' a3388d92-0791-42ad-aaa7-66b388e8ccfa ']' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75506 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 75506 ']' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 75506 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75506 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75506' 00:07:46.775 killing process with pid 75506 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 75506 00:07:46.775 [2024-09-30 14:06:51.388230] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:46.775 [2024-09-30 14:06:51.388317] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:46.775 [2024-09-30 14:06:51.388362] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:46.775 [2024-09-30 14:06:51.388373] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:07:46.775 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 75506 00:07:46.775 [2024-09-30 14:06:51.410181] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:47.034 14:06:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:47.035 00:07:47.035 real 0m4.636s 00:07:47.035 user 0m7.531s 00:07:47.035 sys 0m0.926s 00:07:47.035 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:47.035 14:06:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.035 ************************************ 00:07:47.035 END TEST raid_superblock_test 00:07:47.035 ************************************ 00:07:47.295 14:06:51 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:07:47.295 14:06:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:47.295 14:06:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:47.295 14:06:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:47.295 ************************************ 00:07:47.295 START TEST raid_read_error_test 00:07:47.295 ************************************ 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 read 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.VD4jZju17c 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=75819 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 75819 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 75819 ']' 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:47.295 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:47.295 14:06:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.295 [2024-09-30 14:06:51.823096] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:47.295 [2024-09-30 14:06:51.823298] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75819 ] 00:07:47.555 [2024-09-30 14:06:51.954117] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:47.555 [2024-09-30 14:06:51.981564] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:47.555 [2024-09-30 14:06:52.026300] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:47.555 [2024-09-30 14:06:52.067628] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.555 [2024-09-30 14:06:52.067665] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 BaseBdev1_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 true 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 [2024-09-30 14:06:52.677042] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:48.124 [2024-09-30 14:06:52.677121] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:48.124 [2024-09-30 14:06:52.677143] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:48.124 [2024-09-30 14:06:52.677157] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:48.124 [2024-09-30 14:06:52.679236] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:48.124 [2024-09-30 14:06:52.679349] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:48.124 BaseBdev1 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 BaseBdev2_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 true 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 [2024-09-30 14:06:52.731475] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:48.124 [2024-09-30 14:06:52.731622] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:48.124 [2024-09-30 14:06:52.731675] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:48.124 [2024-09-30 14:06:52.731725] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:48.124 [2024-09-30 14:06:52.734215] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:48.124 [2024-09-30 14:06:52.734301] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:48.124 BaseBdev2 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.124 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.124 [2024-09-30 14:06:52.743444] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:48.124 [2024-09-30 14:06:52.745296] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:48.124 [2024-09-30 14:06:52.745506] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:48.124 [2024-09-30 14:06:52.745567] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:48.124 [2024-09-30 14:06:52.745809] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:48.125 [2024-09-30 14:06:52.745981] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:48.125 [2024-09-30 14:06:52.746030] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:48.125 [2024-09-30 14:06:52.746196] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.125 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.385 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.385 "name": "raid_bdev1", 00:07:48.385 "uuid": "178b85d9-01e9-4977-9c01-8d516bbd540d", 00:07:48.385 "strip_size_kb": 0, 00:07:48.385 "state": "online", 00:07:48.385 "raid_level": "raid1", 00:07:48.385 "superblock": true, 00:07:48.385 "num_base_bdevs": 2, 00:07:48.385 "num_base_bdevs_discovered": 2, 00:07:48.385 "num_base_bdevs_operational": 2, 00:07:48.385 "base_bdevs_list": [ 00:07:48.385 { 00:07:48.385 "name": "BaseBdev1", 00:07:48.385 "uuid": "93ac60eb-c19f-5831-a6cb-3d8d8d183621", 00:07:48.385 "is_configured": true, 00:07:48.385 "data_offset": 2048, 00:07:48.385 "data_size": 63488 00:07:48.385 }, 00:07:48.385 { 00:07:48.385 "name": "BaseBdev2", 00:07:48.385 "uuid": "176bdadd-d1da-585f-93cb-58c69dcebbf1", 00:07:48.385 "is_configured": true, 00:07:48.385 "data_offset": 2048, 00:07:48.385 "data_size": 63488 00:07:48.385 } 00:07:48.385 ] 00:07:48.385 }' 00:07:48.385 14:06:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.385 14:06:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.643 14:06:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:48.643 14:06:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:48.643 [2024-09-30 14:06:53.227017] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.581 "name": "raid_bdev1", 00:07:49.581 "uuid": "178b85d9-01e9-4977-9c01-8d516bbd540d", 00:07:49.581 "strip_size_kb": 0, 00:07:49.581 "state": "online", 00:07:49.581 "raid_level": "raid1", 00:07:49.581 "superblock": true, 00:07:49.581 "num_base_bdevs": 2, 00:07:49.581 "num_base_bdevs_discovered": 2, 00:07:49.581 "num_base_bdevs_operational": 2, 00:07:49.581 "base_bdevs_list": [ 00:07:49.581 { 00:07:49.581 "name": "BaseBdev1", 00:07:49.581 "uuid": "93ac60eb-c19f-5831-a6cb-3d8d8d183621", 00:07:49.581 "is_configured": true, 00:07:49.581 "data_offset": 2048, 00:07:49.581 "data_size": 63488 00:07:49.581 }, 00:07:49.581 { 00:07:49.581 "name": "BaseBdev2", 00:07:49.581 "uuid": "176bdadd-d1da-585f-93cb-58c69dcebbf1", 00:07:49.581 "is_configured": true, 00:07:49.581 "data_offset": 2048, 00:07:49.581 "data_size": 63488 00:07:49.581 } 00:07:49.581 ] 00:07:49.581 }' 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.581 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.150 [2024-09-30 14:06:54.586261] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:50.150 [2024-09-30 14:06:54.586303] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:50.150 [2024-09-30 14:06:54.588748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:50.150 [2024-09-30 14:06:54.588795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:50.150 [2024-09-30 14:06:54.588872] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:50.150 [2024-09-30 14:06:54.588885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:50.150 { 00:07:50.150 "results": [ 00:07:50.150 { 00:07:50.150 "job": "raid_bdev1", 00:07:50.150 "core_mask": "0x1", 00:07:50.150 "workload": "randrw", 00:07:50.150 "percentage": 50, 00:07:50.150 "status": "finished", 00:07:50.150 "queue_depth": 1, 00:07:50.150 "io_size": 131072, 00:07:50.150 "runtime": 1.360068, 00:07:50.150 "iops": 20481.328874732735, 00:07:50.150 "mibps": 2560.166109341592, 00:07:50.150 "io_failed": 0, 00:07:50.150 "io_timeout": 0, 00:07:50.150 "avg_latency_us": 46.40522311877177, 00:07:50.150 "min_latency_us": 21.687336244541484, 00:07:50.150 "max_latency_us": 1409.4532751091704 00:07:50.150 } 00:07:50.150 ], 00:07:50.150 "core_count": 1 00:07:50.150 } 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 75819 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 75819 ']' 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 75819 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75819 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75819' 00:07:50.150 killing process with pid 75819 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 75819 00:07:50.150 [2024-09-30 14:06:54.639240] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:50.150 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 75819 00:07:50.150 [2024-09-30 14:06:54.654781] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.VD4jZju17c 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:50.411 00:07:50.411 real 0m3.172s 00:07:50.411 user 0m3.981s 00:07:50.411 sys 0m0.492s 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:50.411 ************************************ 00:07:50.411 END TEST raid_read_error_test 00:07:50.411 ************************************ 00:07:50.411 14:06:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.411 14:06:54 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:07:50.411 14:06:54 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:50.411 14:06:54 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:50.411 14:06:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:50.411 ************************************ 00:07:50.411 START TEST raid_write_error_test 00:07:50.411 ************************************ 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 write 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Dv42qTWX1f 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=75948 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 75948 00:07:50.411 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 75948 ']' 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:50.411 14:06:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.411 [2024-09-30 14:06:55.063736] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:50.411 [2024-09-30 14:06:55.063869] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75948 ] 00:07:50.671 [2024-09-30 14:06:55.194865] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:50.671 [2024-09-30 14:06:55.223051] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:50.672 [2024-09-30 14:06:55.268244] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:50.672 [2024-09-30 14:06:55.310135] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:50.672 [2024-09-30 14:06:55.310176] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.240 BaseBdev1_malloc 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.240 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 true 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 [2024-09-30 14:06:55.908345] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:51.499 [2024-09-30 14:06:55.908404] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:51.499 [2024-09-30 14:06:55.908422] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:51.499 [2024-09-30 14:06:55.908434] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:51.499 [2024-09-30 14:06:55.910580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:51.499 [2024-09-30 14:06:55.910615] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:51.499 BaseBdev1 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 BaseBdev2_malloc 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 true 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 [2024-09-30 14:06:55.958035] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:51.499 [2024-09-30 14:06:55.958076] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:51.499 [2024-09-30 14:06:55.958091] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:51.499 [2024-09-30 14:06:55.958100] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:51.499 [2024-09-30 14:06:55.960117] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:51.499 [2024-09-30 14:06:55.960218] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:51.499 BaseBdev2 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 [2024-09-30 14:06:55.974049] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:51.499 [2024-09-30 14:06:55.975906] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:51.499 [2024-09-30 14:06:55.976106] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:51.499 [2024-09-30 14:06:55.976155] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:51.499 [2024-09-30 14:06:55.976419] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:07:51.499 [2024-09-30 14:06:55.976598] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:51.499 [2024-09-30 14:06:55.976646] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:07:51.499 [2024-09-30 14:06:55.976800] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.499 14:06:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.499 14:06:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.499 14:06:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.499 "name": "raid_bdev1", 00:07:51.499 "uuid": "51fcdf0f-37b2-4d9d-8832-7eca4df2c33c", 00:07:51.499 "strip_size_kb": 0, 00:07:51.499 "state": "online", 00:07:51.499 "raid_level": "raid1", 00:07:51.499 "superblock": true, 00:07:51.499 "num_base_bdevs": 2, 00:07:51.499 "num_base_bdevs_discovered": 2, 00:07:51.499 "num_base_bdevs_operational": 2, 00:07:51.499 "base_bdevs_list": [ 00:07:51.499 { 00:07:51.499 "name": "BaseBdev1", 00:07:51.499 "uuid": "d10421dd-6992-511b-a0d4-62200a19d53e", 00:07:51.499 "is_configured": true, 00:07:51.499 "data_offset": 2048, 00:07:51.499 "data_size": 63488 00:07:51.499 }, 00:07:51.499 { 00:07:51.499 "name": "BaseBdev2", 00:07:51.499 "uuid": "092547bf-8b4f-5a3d-adf7-7213bc83955b", 00:07:51.500 "is_configured": true, 00:07:51.500 "data_offset": 2048, 00:07:51.500 "data_size": 63488 00:07:51.500 } 00:07:51.500 ] 00:07:51.500 }' 00:07:51.500 14:06:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.500 14:06:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.758 14:06:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:51.758 14:06:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:52.018 [2024-09-30 14:06:56.457629] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.955 [2024-09-30 14:06:57.382048] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:07:52.955 [2024-09-30 14:06:57.382187] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:52.955 [2024-09-30 14:06:57.382427] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005fb0 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:52.955 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.956 "name": "raid_bdev1", 00:07:52.956 "uuid": "51fcdf0f-37b2-4d9d-8832-7eca4df2c33c", 00:07:52.956 "strip_size_kb": 0, 00:07:52.956 "state": "online", 00:07:52.956 "raid_level": "raid1", 00:07:52.956 "superblock": true, 00:07:52.956 "num_base_bdevs": 2, 00:07:52.956 "num_base_bdevs_discovered": 1, 00:07:52.956 "num_base_bdevs_operational": 1, 00:07:52.956 "base_bdevs_list": [ 00:07:52.956 { 00:07:52.956 "name": null, 00:07:52.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.956 "is_configured": false, 00:07:52.956 "data_offset": 0, 00:07:52.956 "data_size": 63488 00:07:52.956 }, 00:07:52.956 { 00:07:52.956 "name": "BaseBdev2", 00:07:52.956 "uuid": "092547bf-8b4f-5a3d-adf7-7213bc83955b", 00:07:52.956 "is_configured": true, 00:07:52.956 "data_offset": 2048, 00:07:52.956 "data_size": 63488 00:07:52.956 } 00:07:52.956 ] 00:07:52.956 }' 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.956 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.215 [2024-09-30 14:06:57.840067] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:53.215 [2024-09-30 14:06:57.840163] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:53.215 [2024-09-30 14:06:57.842563] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:53.215 [2024-09-30 14:06:57.842637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:53.215 [2024-09-30 14:06:57.842711] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:53.215 [2024-09-30 14:06:57.842751] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:07:53.215 { 00:07:53.215 "results": [ 00:07:53.215 { 00:07:53.215 "job": "raid_bdev1", 00:07:53.215 "core_mask": "0x1", 00:07:53.215 "workload": "randrw", 00:07:53.215 "percentage": 50, 00:07:53.215 "status": "finished", 00:07:53.215 "queue_depth": 1, 00:07:53.215 "io_size": 131072, 00:07:53.215 "runtime": 1.383324, 00:07:53.215 "iops": 24099.19874158187, 00:07:53.215 "mibps": 3012.3998426977337, 00:07:53.215 "io_failed": 0, 00:07:53.215 "io_timeout": 0, 00:07:53.215 "avg_latency_us": 39.064996404980604, 00:07:53.215 "min_latency_us": 20.90480349344978, 00:07:53.215 "max_latency_us": 1359.3711790393013 00:07:53.215 } 00:07:53.215 ], 00:07:53.215 "core_count": 1 00:07:53.215 } 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 75948 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 75948 ']' 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 75948 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:53.215 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75948 00:07:53.475 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:53.475 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:53.475 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75948' 00:07:53.475 killing process with pid 75948 00:07:53.475 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 75948 00:07:53.475 [2024-09-30 14:06:57.884333] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:53.475 14:06:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 75948 00:07:53.475 [2024-09-30 14:06:57.899100] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:53.475 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Dv42qTWX1f 00:07:53.475 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:53.735 00:07:53.735 real 0m3.182s 00:07:53.735 user 0m3.969s 00:07:53.735 sys 0m0.525s 00:07:53.735 ************************************ 00:07:53.735 END TEST raid_write_error_test 00:07:53.735 ************************************ 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:53.735 14:06:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.735 14:06:58 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:53.735 14:06:58 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:53.735 14:06:58 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:07:53.735 14:06:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:53.735 14:06:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:53.735 14:06:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:53.735 ************************************ 00:07:53.735 START TEST raid_state_function_test 00:07:53.735 ************************************ 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 false 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76081 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76081' 00:07:53.735 Process raid pid: 76081 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76081 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 76081 ']' 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:53.735 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:53.735 14:06:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.735 [2024-09-30 14:06:58.296581] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:07:53.735 [2024-09-30 14:06:58.296796] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:53.995 [2024-09-30 14:06:58.428765] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:07:53.995 [2024-09-30 14:06:58.456511] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:53.995 [2024-09-30 14:06:58.501770] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:53.995 [2024-09-30 14:06:58.543041] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:53.995 [2024-09-30 14:06:58.543154] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.562 [2024-09-30 14:06:59.151845] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:54.562 [2024-09-30 14:06:59.151895] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:54.562 [2024-09-30 14:06:59.151908] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:54.562 [2024-09-30 14:06:59.151916] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:54.562 [2024-09-30 14:06:59.151927] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:54.562 [2024-09-30 14:06:59.151935] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.562 "name": "Existed_Raid", 00:07:54.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.562 "strip_size_kb": 64, 00:07:54.562 "state": "configuring", 00:07:54.562 "raid_level": "raid0", 00:07:54.562 "superblock": false, 00:07:54.562 "num_base_bdevs": 3, 00:07:54.562 "num_base_bdevs_discovered": 0, 00:07:54.562 "num_base_bdevs_operational": 3, 00:07:54.562 "base_bdevs_list": [ 00:07:54.562 { 00:07:54.562 "name": "BaseBdev1", 00:07:54.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.562 "is_configured": false, 00:07:54.562 "data_offset": 0, 00:07:54.562 "data_size": 0 00:07:54.562 }, 00:07:54.562 { 00:07:54.562 "name": "BaseBdev2", 00:07:54.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.562 "is_configured": false, 00:07:54.562 "data_offset": 0, 00:07:54.562 "data_size": 0 00:07:54.562 }, 00:07:54.562 { 00:07:54.562 "name": "BaseBdev3", 00:07:54.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.562 "is_configured": false, 00:07:54.562 "data_offset": 0, 00:07:54.562 "data_size": 0 00:07:54.562 } 00:07:54.562 ] 00:07:54.562 }' 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.562 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 [2024-09-30 14:06:59.531106] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:55.130 [2024-09-30 14:06:59.531183] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 [2024-09-30 14:06:59.543093] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:55.130 [2024-09-30 14:06:59.543164] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:55.130 [2024-09-30 14:06:59.543208] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:55.130 [2024-09-30 14:06:59.543228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:55.130 [2024-09-30 14:06:59.543247] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:55.130 [2024-09-30 14:06:59.543266] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 [2024-09-30 14:06:59.563693] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:55.130 BaseBdev1 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.130 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.130 [ 00:07:55.130 { 00:07:55.130 "name": "BaseBdev1", 00:07:55.130 "aliases": [ 00:07:55.130 "222278cf-53fd-4f6d-8e4e-8fd6f817c92a" 00:07:55.130 ], 00:07:55.130 "product_name": "Malloc disk", 00:07:55.130 "block_size": 512, 00:07:55.130 "num_blocks": 65536, 00:07:55.131 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:55.131 "assigned_rate_limits": { 00:07:55.131 "rw_ios_per_sec": 0, 00:07:55.131 "rw_mbytes_per_sec": 0, 00:07:55.131 "r_mbytes_per_sec": 0, 00:07:55.131 "w_mbytes_per_sec": 0 00:07:55.131 }, 00:07:55.131 "claimed": true, 00:07:55.131 "claim_type": "exclusive_write", 00:07:55.131 "zoned": false, 00:07:55.131 "supported_io_types": { 00:07:55.131 "read": true, 00:07:55.131 "write": true, 00:07:55.131 "unmap": true, 00:07:55.131 "flush": true, 00:07:55.131 "reset": true, 00:07:55.131 "nvme_admin": false, 00:07:55.131 "nvme_io": false, 00:07:55.131 "nvme_io_md": false, 00:07:55.131 "write_zeroes": true, 00:07:55.131 "zcopy": true, 00:07:55.131 "get_zone_info": false, 00:07:55.131 "zone_management": false, 00:07:55.131 "zone_append": false, 00:07:55.131 "compare": false, 00:07:55.131 "compare_and_write": false, 00:07:55.131 "abort": true, 00:07:55.131 "seek_hole": false, 00:07:55.131 "seek_data": false, 00:07:55.131 "copy": true, 00:07:55.131 "nvme_iov_md": false 00:07:55.131 }, 00:07:55.131 "memory_domains": [ 00:07:55.131 { 00:07:55.131 "dma_device_id": "system", 00:07:55.131 "dma_device_type": 1 00:07:55.131 }, 00:07:55.131 { 00:07:55.131 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.131 "dma_device_type": 2 00:07:55.131 } 00:07:55.131 ], 00:07:55.131 "driver_specific": {} 00:07:55.131 } 00:07:55.131 ] 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.131 "name": "Existed_Raid", 00:07:55.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.131 "strip_size_kb": 64, 00:07:55.131 "state": "configuring", 00:07:55.131 "raid_level": "raid0", 00:07:55.131 "superblock": false, 00:07:55.131 "num_base_bdevs": 3, 00:07:55.131 "num_base_bdevs_discovered": 1, 00:07:55.131 "num_base_bdevs_operational": 3, 00:07:55.131 "base_bdevs_list": [ 00:07:55.131 { 00:07:55.131 "name": "BaseBdev1", 00:07:55.131 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:55.131 "is_configured": true, 00:07:55.131 "data_offset": 0, 00:07:55.131 "data_size": 65536 00:07:55.131 }, 00:07:55.131 { 00:07:55.131 "name": "BaseBdev2", 00:07:55.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.131 "is_configured": false, 00:07:55.131 "data_offset": 0, 00:07:55.131 "data_size": 0 00:07:55.131 }, 00:07:55.131 { 00:07:55.131 "name": "BaseBdev3", 00:07:55.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.131 "is_configured": false, 00:07:55.131 "data_offset": 0, 00:07:55.131 "data_size": 0 00:07:55.131 } 00:07:55.131 ] 00:07:55.131 }' 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.131 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.390 [2024-09-30 14:06:59.983035] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:55.390 [2024-09-30 14:06:59.983122] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.390 [2024-09-30 14:06:59.995012] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:55.390 [2024-09-30 14:06:59.996843] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:55.390 [2024-09-30 14:06:59.996881] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:55.390 [2024-09-30 14:06:59.996893] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:55.390 [2024-09-30 14:06:59.996901] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:55.390 14:06:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.390 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.649 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.649 "name": "Existed_Raid", 00:07:55.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.649 "strip_size_kb": 64, 00:07:55.649 "state": "configuring", 00:07:55.649 "raid_level": "raid0", 00:07:55.649 "superblock": false, 00:07:55.649 "num_base_bdevs": 3, 00:07:55.649 "num_base_bdevs_discovered": 1, 00:07:55.649 "num_base_bdevs_operational": 3, 00:07:55.649 "base_bdevs_list": [ 00:07:55.649 { 00:07:55.649 "name": "BaseBdev1", 00:07:55.649 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:55.649 "is_configured": true, 00:07:55.649 "data_offset": 0, 00:07:55.649 "data_size": 65536 00:07:55.649 }, 00:07:55.649 { 00:07:55.649 "name": "BaseBdev2", 00:07:55.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.649 "is_configured": false, 00:07:55.649 "data_offset": 0, 00:07:55.649 "data_size": 0 00:07:55.649 }, 00:07:55.649 { 00:07:55.649 "name": "BaseBdev3", 00:07:55.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.649 "is_configured": false, 00:07:55.649 "data_offset": 0, 00:07:55.649 "data_size": 0 00:07:55.649 } 00:07:55.649 ] 00:07:55.649 }' 00:07:55.649 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.649 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.909 [2024-09-30 14:07:00.437002] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:55.909 BaseBdev2 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.909 [ 00:07:55.909 { 00:07:55.909 "name": "BaseBdev2", 00:07:55.909 "aliases": [ 00:07:55.909 "36b0e950-805f-4f34-84a7-f0c43f75a4d7" 00:07:55.909 ], 00:07:55.909 "product_name": "Malloc disk", 00:07:55.909 "block_size": 512, 00:07:55.909 "num_blocks": 65536, 00:07:55.909 "uuid": "36b0e950-805f-4f34-84a7-f0c43f75a4d7", 00:07:55.909 "assigned_rate_limits": { 00:07:55.909 "rw_ios_per_sec": 0, 00:07:55.909 "rw_mbytes_per_sec": 0, 00:07:55.909 "r_mbytes_per_sec": 0, 00:07:55.909 "w_mbytes_per_sec": 0 00:07:55.909 }, 00:07:55.909 "claimed": true, 00:07:55.909 "claim_type": "exclusive_write", 00:07:55.909 "zoned": false, 00:07:55.909 "supported_io_types": { 00:07:55.909 "read": true, 00:07:55.909 "write": true, 00:07:55.909 "unmap": true, 00:07:55.909 "flush": true, 00:07:55.909 "reset": true, 00:07:55.909 "nvme_admin": false, 00:07:55.909 "nvme_io": false, 00:07:55.909 "nvme_io_md": false, 00:07:55.909 "write_zeroes": true, 00:07:55.909 "zcopy": true, 00:07:55.909 "get_zone_info": false, 00:07:55.909 "zone_management": false, 00:07:55.909 "zone_append": false, 00:07:55.909 "compare": false, 00:07:55.909 "compare_and_write": false, 00:07:55.909 "abort": true, 00:07:55.909 "seek_hole": false, 00:07:55.909 "seek_data": false, 00:07:55.909 "copy": true, 00:07:55.909 "nvme_iov_md": false 00:07:55.909 }, 00:07:55.909 "memory_domains": [ 00:07:55.909 { 00:07:55.909 "dma_device_id": "system", 00:07:55.909 "dma_device_type": 1 00:07:55.909 }, 00:07:55.909 { 00:07:55.909 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.909 "dma_device_type": 2 00:07:55.909 } 00:07:55.909 ], 00:07:55.909 "driver_specific": {} 00:07:55.909 } 00:07:55.909 ] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.909 "name": "Existed_Raid", 00:07:55.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.909 "strip_size_kb": 64, 00:07:55.909 "state": "configuring", 00:07:55.909 "raid_level": "raid0", 00:07:55.909 "superblock": false, 00:07:55.909 "num_base_bdevs": 3, 00:07:55.909 "num_base_bdevs_discovered": 2, 00:07:55.909 "num_base_bdevs_operational": 3, 00:07:55.909 "base_bdevs_list": [ 00:07:55.909 { 00:07:55.909 "name": "BaseBdev1", 00:07:55.909 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:55.909 "is_configured": true, 00:07:55.909 "data_offset": 0, 00:07:55.909 "data_size": 65536 00:07:55.909 }, 00:07:55.909 { 00:07:55.909 "name": "BaseBdev2", 00:07:55.909 "uuid": "36b0e950-805f-4f34-84a7-f0c43f75a4d7", 00:07:55.909 "is_configured": true, 00:07:55.909 "data_offset": 0, 00:07:55.909 "data_size": 65536 00:07:55.909 }, 00:07:55.909 { 00:07:55.909 "name": "BaseBdev3", 00:07:55.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.909 "is_configured": false, 00:07:55.909 "data_offset": 0, 00:07:55.909 "data_size": 0 00:07:55.909 } 00:07:55.909 ] 00:07:55.909 }' 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.909 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.477 [2024-09-30 14:07:00.867054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:56.477 [2024-09-30 14:07:00.867096] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:07:56.477 [2024-09-30 14:07:00.867112] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:56.477 [2024-09-30 14:07:00.867376] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:56.477 [2024-09-30 14:07:00.867530] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:07:56.477 [2024-09-30 14:07:00.867548] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:07:56.477 [2024-09-30 14:07:00.867760] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:56.477 BaseBdev3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.477 [ 00:07:56.477 { 00:07:56.477 "name": "BaseBdev3", 00:07:56.477 "aliases": [ 00:07:56.477 "bd0eb3c2-820a-41ca-8ff2-994d5f622c48" 00:07:56.477 ], 00:07:56.477 "product_name": "Malloc disk", 00:07:56.477 "block_size": 512, 00:07:56.477 "num_blocks": 65536, 00:07:56.477 "uuid": "bd0eb3c2-820a-41ca-8ff2-994d5f622c48", 00:07:56.477 "assigned_rate_limits": { 00:07:56.477 "rw_ios_per_sec": 0, 00:07:56.477 "rw_mbytes_per_sec": 0, 00:07:56.477 "r_mbytes_per_sec": 0, 00:07:56.477 "w_mbytes_per_sec": 0 00:07:56.477 }, 00:07:56.477 "claimed": true, 00:07:56.477 "claim_type": "exclusive_write", 00:07:56.477 "zoned": false, 00:07:56.477 "supported_io_types": { 00:07:56.477 "read": true, 00:07:56.477 "write": true, 00:07:56.477 "unmap": true, 00:07:56.477 "flush": true, 00:07:56.477 "reset": true, 00:07:56.477 "nvme_admin": false, 00:07:56.477 "nvme_io": false, 00:07:56.477 "nvme_io_md": false, 00:07:56.477 "write_zeroes": true, 00:07:56.477 "zcopy": true, 00:07:56.477 "get_zone_info": false, 00:07:56.477 "zone_management": false, 00:07:56.477 "zone_append": false, 00:07:56.477 "compare": false, 00:07:56.477 "compare_and_write": false, 00:07:56.477 "abort": true, 00:07:56.477 "seek_hole": false, 00:07:56.477 "seek_data": false, 00:07:56.477 "copy": true, 00:07:56.477 "nvme_iov_md": false 00:07:56.477 }, 00:07:56.477 "memory_domains": [ 00:07:56.477 { 00:07:56.477 "dma_device_id": "system", 00:07:56.477 "dma_device_type": 1 00:07:56.477 }, 00:07:56.477 { 00:07:56.477 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.477 "dma_device_type": 2 00:07:56.477 } 00:07:56.477 ], 00:07:56.477 "driver_specific": {} 00:07:56.477 } 00:07:56.477 ] 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.477 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.478 "name": "Existed_Raid", 00:07:56.478 "uuid": "87eec6dd-8275-43d8-b8e4-750221800bb1", 00:07:56.478 "strip_size_kb": 64, 00:07:56.478 "state": "online", 00:07:56.478 "raid_level": "raid0", 00:07:56.478 "superblock": false, 00:07:56.478 "num_base_bdevs": 3, 00:07:56.478 "num_base_bdevs_discovered": 3, 00:07:56.478 "num_base_bdevs_operational": 3, 00:07:56.478 "base_bdevs_list": [ 00:07:56.478 { 00:07:56.478 "name": "BaseBdev1", 00:07:56.478 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:56.478 "is_configured": true, 00:07:56.478 "data_offset": 0, 00:07:56.478 "data_size": 65536 00:07:56.478 }, 00:07:56.478 { 00:07:56.478 "name": "BaseBdev2", 00:07:56.478 "uuid": "36b0e950-805f-4f34-84a7-f0c43f75a4d7", 00:07:56.478 "is_configured": true, 00:07:56.478 "data_offset": 0, 00:07:56.478 "data_size": 65536 00:07:56.478 }, 00:07:56.478 { 00:07:56.478 "name": "BaseBdev3", 00:07:56.478 "uuid": "bd0eb3c2-820a-41ca-8ff2-994d5f622c48", 00:07:56.478 "is_configured": true, 00:07:56.478 "data_offset": 0, 00:07:56.478 "data_size": 65536 00:07:56.478 } 00:07:56.478 ] 00:07:56.478 }' 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.478 14:07:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.738 [2024-09-30 14:07:01.294725] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:56.738 "name": "Existed_Raid", 00:07:56.738 "aliases": [ 00:07:56.738 "87eec6dd-8275-43d8-b8e4-750221800bb1" 00:07:56.738 ], 00:07:56.738 "product_name": "Raid Volume", 00:07:56.738 "block_size": 512, 00:07:56.738 "num_blocks": 196608, 00:07:56.738 "uuid": "87eec6dd-8275-43d8-b8e4-750221800bb1", 00:07:56.738 "assigned_rate_limits": { 00:07:56.738 "rw_ios_per_sec": 0, 00:07:56.738 "rw_mbytes_per_sec": 0, 00:07:56.738 "r_mbytes_per_sec": 0, 00:07:56.738 "w_mbytes_per_sec": 0 00:07:56.738 }, 00:07:56.738 "claimed": false, 00:07:56.738 "zoned": false, 00:07:56.738 "supported_io_types": { 00:07:56.738 "read": true, 00:07:56.738 "write": true, 00:07:56.738 "unmap": true, 00:07:56.738 "flush": true, 00:07:56.738 "reset": true, 00:07:56.738 "nvme_admin": false, 00:07:56.738 "nvme_io": false, 00:07:56.738 "nvme_io_md": false, 00:07:56.738 "write_zeroes": true, 00:07:56.738 "zcopy": false, 00:07:56.738 "get_zone_info": false, 00:07:56.738 "zone_management": false, 00:07:56.738 "zone_append": false, 00:07:56.738 "compare": false, 00:07:56.738 "compare_and_write": false, 00:07:56.738 "abort": false, 00:07:56.738 "seek_hole": false, 00:07:56.738 "seek_data": false, 00:07:56.738 "copy": false, 00:07:56.738 "nvme_iov_md": false 00:07:56.738 }, 00:07:56.738 "memory_domains": [ 00:07:56.738 { 00:07:56.738 "dma_device_id": "system", 00:07:56.738 "dma_device_type": 1 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.738 "dma_device_type": 2 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "dma_device_id": "system", 00:07:56.738 "dma_device_type": 1 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.738 "dma_device_type": 2 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "dma_device_id": "system", 00:07:56.738 "dma_device_type": 1 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.738 "dma_device_type": 2 00:07:56.738 } 00:07:56.738 ], 00:07:56.738 "driver_specific": { 00:07:56.738 "raid": { 00:07:56.738 "uuid": "87eec6dd-8275-43d8-b8e4-750221800bb1", 00:07:56.738 "strip_size_kb": 64, 00:07:56.738 "state": "online", 00:07:56.738 "raid_level": "raid0", 00:07:56.738 "superblock": false, 00:07:56.738 "num_base_bdevs": 3, 00:07:56.738 "num_base_bdevs_discovered": 3, 00:07:56.738 "num_base_bdevs_operational": 3, 00:07:56.738 "base_bdevs_list": [ 00:07:56.738 { 00:07:56.738 "name": "BaseBdev1", 00:07:56.738 "uuid": "222278cf-53fd-4f6d-8e4e-8fd6f817c92a", 00:07:56.738 "is_configured": true, 00:07:56.738 "data_offset": 0, 00:07:56.738 "data_size": 65536 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "name": "BaseBdev2", 00:07:56.738 "uuid": "36b0e950-805f-4f34-84a7-f0c43f75a4d7", 00:07:56.738 "is_configured": true, 00:07:56.738 "data_offset": 0, 00:07:56.738 "data_size": 65536 00:07:56.738 }, 00:07:56.738 { 00:07:56.738 "name": "BaseBdev3", 00:07:56.738 "uuid": "bd0eb3c2-820a-41ca-8ff2-994d5f622c48", 00:07:56.738 "is_configured": true, 00:07:56.738 "data_offset": 0, 00:07:56.738 "data_size": 65536 00:07:56.738 } 00:07:56.738 ] 00:07:56.738 } 00:07:56.738 } 00:07:56.738 }' 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:56.738 BaseBdev2 00:07:56.738 BaseBdev3' 00:07:56.738 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.998 [2024-09-30 14:07:01.530041] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:56.998 [2024-09-30 14:07:01.530069] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:56.998 [2024-09-30 14:07:01.530121] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.998 "name": "Existed_Raid", 00:07:56.998 "uuid": "87eec6dd-8275-43d8-b8e4-750221800bb1", 00:07:56.998 "strip_size_kb": 64, 00:07:56.998 "state": "offline", 00:07:56.998 "raid_level": "raid0", 00:07:56.998 "superblock": false, 00:07:56.998 "num_base_bdevs": 3, 00:07:56.998 "num_base_bdevs_discovered": 2, 00:07:56.998 "num_base_bdevs_operational": 2, 00:07:56.998 "base_bdevs_list": [ 00:07:56.998 { 00:07:56.998 "name": null, 00:07:56.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.998 "is_configured": false, 00:07:56.998 "data_offset": 0, 00:07:56.998 "data_size": 65536 00:07:56.998 }, 00:07:56.998 { 00:07:56.998 "name": "BaseBdev2", 00:07:56.998 "uuid": "36b0e950-805f-4f34-84a7-f0c43f75a4d7", 00:07:56.998 "is_configured": true, 00:07:56.998 "data_offset": 0, 00:07:56.998 "data_size": 65536 00:07:56.998 }, 00:07:56.998 { 00:07:56.998 "name": "BaseBdev3", 00:07:56.998 "uuid": "bd0eb3c2-820a-41ca-8ff2-994d5f622c48", 00:07:56.998 "is_configured": true, 00:07:56.998 "data_offset": 0, 00:07:56.998 "data_size": 65536 00:07:56.998 } 00:07:56.998 ] 00:07:56.998 }' 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.998 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.567 [2024-09-30 14:07:01.980410] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.567 14:07:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.567 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.567 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:57.567 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:57.567 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:57.567 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 [2024-09-30 14:07:02.039206] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:57.568 [2024-09-30 14:07:02.039255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 BaseBdev2 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 [ 00:07:57.568 { 00:07:57.568 "name": "BaseBdev2", 00:07:57.568 "aliases": [ 00:07:57.568 "47ab4145-b91a-4f94-871c-477931c4c7b3" 00:07:57.568 ], 00:07:57.568 "product_name": "Malloc disk", 00:07:57.568 "block_size": 512, 00:07:57.568 "num_blocks": 65536, 00:07:57.568 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:57.568 "assigned_rate_limits": { 00:07:57.568 "rw_ios_per_sec": 0, 00:07:57.568 "rw_mbytes_per_sec": 0, 00:07:57.568 "r_mbytes_per_sec": 0, 00:07:57.568 "w_mbytes_per_sec": 0 00:07:57.568 }, 00:07:57.568 "claimed": false, 00:07:57.568 "zoned": false, 00:07:57.568 "supported_io_types": { 00:07:57.568 "read": true, 00:07:57.568 "write": true, 00:07:57.568 "unmap": true, 00:07:57.568 "flush": true, 00:07:57.568 "reset": true, 00:07:57.568 "nvme_admin": false, 00:07:57.568 "nvme_io": false, 00:07:57.568 "nvme_io_md": false, 00:07:57.568 "write_zeroes": true, 00:07:57.568 "zcopy": true, 00:07:57.568 "get_zone_info": false, 00:07:57.568 "zone_management": false, 00:07:57.568 "zone_append": false, 00:07:57.568 "compare": false, 00:07:57.568 "compare_and_write": false, 00:07:57.568 "abort": true, 00:07:57.568 "seek_hole": false, 00:07:57.568 "seek_data": false, 00:07:57.568 "copy": true, 00:07:57.568 "nvme_iov_md": false 00:07:57.568 }, 00:07:57.568 "memory_domains": [ 00:07:57.568 { 00:07:57.568 "dma_device_id": "system", 00:07:57.568 "dma_device_type": 1 00:07:57.568 }, 00:07:57.568 { 00:07:57.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.568 "dma_device_type": 2 00:07:57.568 } 00:07:57.568 ], 00:07:57.568 "driver_specific": {} 00:07:57.568 } 00:07:57.568 ] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 BaseBdev3 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 [ 00:07:57.568 { 00:07:57.568 "name": "BaseBdev3", 00:07:57.568 "aliases": [ 00:07:57.568 "1959c96f-aeb7-4527-8cc6-46563210c6b0" 00:07:57.568 ], 00:07:57.568 "product_name": "Malloc disk", 00:07:57.568 "block_size": 512, 00:07:57.568 "num_blocks": 65536, 00:07:57.568 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:57.568 "assigned_rate_limits": { 00:07:57.568 "rw_ios_per_sec": 0, 00:07:57.568 "rw_mbytes_per_sec": 0, 00:07:57.568 "r_mbytes_per_sec": 0, 00:07:57.568 "w_mbytes_per_sec": 0 00:07:57.568 }, 00:07:57.568 "claimed": false, 00:07:57.568 "zoned": false, 00:07:57.568 "supported_io_types": { 00:07:57.568 "read": true, 00:07:57.568 "write": true, 00:07:57.568 "unmap": true, 00:07:57.568 "flush": true, 00:07:57.568 "reset": true, 00:07:57.568 "nvme_admin": false, 00:07:57.568 "nvme_io": false, 00:07:57.568 "nvme_io_md": false, 00:07:57.568 "write_zeroes": true, 00:07:57.568 "zcopy": true, 00:07:57.568 "get_zone_info": false, 00:07:57.568 "zone_management": false, 00:07:57.568 "zone_append": false, 00:07:57.568 "compare": false, 00:07:57.568 "compare_and_write": false, 00:07:57.568 "abort": true, 00:07:57.568 "seek_hole": false, 00:07:57.568 "seek_data": false, 00:07:57.568 "copy": true, 00:07:57.568 "nvme_iov_md": false 00:07:57.568 }, 00:07:57.568 "memory_domains": [ 00:07:57.568 { 00:07:57.568 "dma_device_id": "system", 00:07:57.568 "dma_device_type": 1 00:07:57.568 }, 00:07:57.568 { 00:07:57.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.568 "dma_device_type": 2 00:07:57.568 } 00:07:57.568 ], 00:07:57.568 "driver_specific": {} 00:07:57.568 } 00:07:57.568 ] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.568 [2024-09-30 14:07:02.213875] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:57.568 [2024-09-30 14:07:02.213967] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:57.568 [2024-09-30 14:07:02.214011] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:57.568 [2024-09-30 14:07:02.215822] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:57.568 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.569 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.828 "name": "Existed_Raid", 00:07:57.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.828 "strip_size_kb": 64, 00:07:57.828 "state": "configuring", 00:07:57.828 "raid_level": "raid0", 00:07:57.828 "superblock": false, 00:07:57.828 "num_base_bdevs": 3, 00:07:57.828 "num_base_bdevs_discovered": 2, 00:07:57.828 "num_base_bdevs_operational": 3, 00:07:57.828 "base_bdevs_list": [ 00:07:57.828 { 00:07:57.828 "name": "BaseBdev1", 00:07:57.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.828 "is_configured": false, 00:07:57.828 "data_offset": 0, 00:07:57.828 "data_size": 0 00:07:57.828 }, 00:07:57.828 { 00:07:57.828 "name": "BaseBdev2", 00:07:57.828 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:57.828 "is_configured": true, 00:07:57.828 "data_offset": 0, 00:07:57.828 "data_size": 65536 00:07:57.828 }, 00:07:57.828 { 00:07:57.828 "name": "BaseBdev3", 00:07:57.828 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:57.828 "is_configured": true, 00:07:57.828 "data_offset": 0, 00:07:57.828 "data_size": 65536 00:07:57.828 } 00:07:57.828 ] 00:07:57.828 }' 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.828 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.087 [2024-09-30 14:07:02.645138] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.087 "name": "Existed_Raid", 00:07:58.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.087 "strip_size_kb": 64, 00:07:58.087 "state": "configuring", 00:07:58.087 "raid_level": "raid0", 00:07:58.087 "superblock": false, 00:07:58.087 "num_base_bdevs": 3, 00:07:58.087 "num_base_bdevs_discovered": 1, 00:07:58.087 "num_base_bdevs_operational": 3, 00:07:58.087 "base_bdevs_list": [ 00:07:58.087 { 00:07:58.087 "name": "BaseBdev1", 00:07:58.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.087 "is_configured": false, 00:07:58.087 "data_offset": 0, 00:07:58.087 "data_size": 0 00:07:58.087 }, 00:07:58.087 { 00:07:58.087 "name": null, 00:07:58.087 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:58.087 "is_configured": false, 00:07:58.087 "data_offset": 0, 00:07:58.087 "data_size": 65536 00:07:58.087 }, 00:07:58.087 { 00:07:58.087 "name": "BaseBdev3", 00:07:58.087 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:58.087 "is_configured": true, 00:07:58.087 "data_offset": 0, 00:07:58.087 "data_size": 65536 00:07:58.087 } 00:07:58.087 ] 00:07:58.087 }' 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.087 14:07:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.660 [2024-09-30 14:07:03.115114] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:58.660 BaseBdev1 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:58.660 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.661 [ 00:07:58.661 { 00:07:58.661 "name": "BaseBdev1", 00:07:58.661 "aliases": [ 00:07:58.661 "c8c2af9f-8219-40c1-937f-a15d7258aead" 00:07:58.661 ], 00:07:58.661 "product_name": "Malloc disk", 00:07:58.661 "block_size": 512, 00:07:58.661 "num_blocks": 65536, 00:07:58.661 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:07:58.661 "assigned_rate_limits": { 00:07:58.661 "rw_ios_per_sec": 0, 00:07:58.661 "rw_mbytes_per_sec": 0, 00:07:58.661 "r_mbytes_per_sec": 0, 00:07:58.661 "w_mbytes_per_sec": 0 00:07:58.661 }, 00:07:58.661 "claimed": true, 00:07:58.661 "claim_type": "exclusive_write", 00:07:58.661 "zoned": false, 00:07:58.661 "supported_io_types": { 00:07:58.661 "read": true, 00:07:58.661 "write": true, 00:07:58.661 "unmap": true, 00:07:58.661 "flush": true, 00:07:58.661 "reset": true, 00:07:58.661 "nvme_admin": false, 00:07:58.661 "nvme_io": false, 00:07:58.661 "nvme_io_md": false, 00:07:58.661 "write_zeroes": true, 00:07:58.661 "zcopy": true, 00:07:58.661 "get_zone_info": false, 00:07:58.661 "zone_management": false, 00:07:58.661 "zone_append": false, 00:07:58.661 "compare": false, 00:07:58.661 "compare_and_write": false, 00:07:58.661 "abort": true, 00:07:58.661 "seek_hole": false, 00:07:58.661 "seek_data": false, 00:07:58.661 "copy": true, 00:07:58.661 "nvme_iov_md": false 00:07:58.661 }, 00:07:58.661 "memory_domains": [ 00:07:58.661 { 00:07:58.661 "dma_device_id": "system", 00:07:58.661 "dma_device_type": 1 00:07:58.661 }, 00:07:58.661 { 00:07:58.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.661 "dma_device_type": 2 00:07:58.661 } 00:07:58.661 ], 00:07:58.661 "driver_specific": {} 00:07:58.661 } 00:07:58.661 ] 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.661 "name": "Existed_Raid", 00:07:58.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.661 "strip_size_kb": 64, 00:07:58.661 "state": "configuring", 00:07:58.661 "raid_level": "raid0", 00:07:58.661 "superblock": false, 00:07:58.661 "num_base_bdevs": 3, 00:07:58.661 "num_base_bdevs_discovered": 2, 00:07:58.661 "num_base_bdevs_operational": 3, 00:07:58.661 "base_bdevs_list": [ 00:07:58.661 { 00:07:58.661 "name": "BaseBdev1", 00:07:58.661 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:07:58.661 "is_configured": true, 00:07:58.661 "data_offset": 0, 00:07:58.661 "data_size": 65536 00:07:58.661 }, 00:07:58.661 { 00:07:58.661 "name": null, 00:07:58.661 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:58.661 "is_configured": false, 00:07:58.661 "data_offset": 0, 00:07:58.661 "data_size": 65536 00:07:58.661 }, 00:07:58.661 { 00:07:58.661 "name": "BaseBdev3", 00:07:58.661 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:58.661 "is_configured": true, 00:07:58.661 "data_offset": 0, 00:07:58.661 "data_size": 65536 00:07:58.661 } 00:07:58.661 ] 00:07:58.661 }' 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.661 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.953 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.953 [2024-09-30 14:07:03.606337] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.213 "name": "Existed_Raid", 00:07:59.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.213 "strip_size_kb": 64, 00:07:59.213 "state": "configuring", 00:07:59.213 "raid_level": "raid0", 00:07:59.213 "superblock": false, 00:07:59.213 "num_base_bdevs": 3, 00:07:59.213 "num_base_bdevs_discovered": 1, 00:07:59.213 "num_base_bdevs_operational": 3, 00:07:59.213 "base_bdevs_list": [ 00:07:59.213 { 00:07:59.213 "name": "BaseBdev1", 00:07:59.213 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:07:59.213 "is_configured": true, 00:07:59.213 "data_offset": 0, 00:07:59.213 "data_size": 65536 00:07:59.213 }, 00:07:59.213 { 00:07:59.213 "name": null, 00:07:59.213 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:59.213 "is_configured": false, 00:07:59.213 "data_offset": 0, 00:07:59.213 "data_size": 65536 00:07:59.213 }, 00:07:59.213 { 00:07:59.213 "name": null, 00:07:59.213 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:59.213 "is_configured": false, 00:07:59.213 "data_offset": 0, 00:07:59.213 "data_size": 65536 00:07:59.213 } 00:07:59.213 ] 00:07:59.213 }' 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.213 14:07:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.473 [2024-09-30 14:07:04.109515] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.473 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.732 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.732 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.732 "name": "Existed_Raid", 00:07:59.732 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.732 "strip_size_kb": 64, 00:07:59.732 "state": "configuring", 00:07:59.732 "raid_level": "raid0", 00:07:59.732 "superblock": false, 00:07:59.732 "num_base_bdevs": 3, 00:07:59.732 "num_base_bdevs_discovered": 2, 00:07:59.732 "num_base_bdevs_operational": 3, 00:07:59.732 "base_bdevs_list": [ 00:07:59.732 { 00:07:59.732 "name": "BaseBdev1", 00:07:59.732 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:07:59.732 "is_configured": true, 00:07:59.732 "data_offset": 0, 00:07:59.732 "data_size": 65536 00:07:59.732 }, 00:07:59.732 { 00:07:59.732 "name": null, 00:07:59.732 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:07:59.732 "is_configured": false, 00:07:59.732 "data_offset": 0, 00:07:59.732 "data_size": 65536 00:07:59.732 }, 00:07:59.732 { 00:07:59.732 "name": "BaseBdev3", 00:07:59.732 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:07:59.732 "is_configured": true, 00:07:59.732 "data_offset": 0, 00:07:59.732 "data_size": 65536 00:07:59.732 } 00:07:59.732 ] 00:07:59.732 }' 00:07:59.732 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.732 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.991 [2024-09-30 14:07:04.588692] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.991 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.250 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.250 "name": "Existed_Raid", 00:08:00.250 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:00.250 "strip_size_kb": 64, 00:08:00.250 "state": "configuring", 00:08:00.250 "raid_level": "raid0", 00:08:00.250 "superblock": false, 00:08:00.250 "num_base_bdevs": 3, 00:08:00.250 "num_base_bdevs_discovered": 1, 00:08:00.250 "num_base_bdevs_operational": 3, 00:08:00.250 "base_bdevs_list": [ 00:08:00.250 { 00:08:00.250 "name": null, 00:08:00.250 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:08:00.250 "is_configured": false, 00:08:00.250 "data_offset": 0, 00:08:00.250 "data_size": 65536 00:08:00.250 }, 00:08:00.250 { 00:08:00.250 "name": null, 00:08:00.250 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:08:00.250 "is_configured": false, 00:08:00.250 "data_offset": 0, 00:08:00.250 "data_size": 65536 00:08:00.250 }, 00:08:00.250 { 00:08:00.250 "name": "BaseBdev3", 00:08:00.250 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:08:00.250 "is_configured": true, 00:08:00.250 "data_offset": 0, 00:08:00.250 "data_size": 65536 00:08:00.250 } 00:08:00.250 ] 00:08:00.250 }' 00:08:00.250 14:07:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.250 14:07:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.510 [2024-09-30 14:07:05.094208] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.510 "name": "Existed_Raid", 00:08:00.510 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:00.510 "strip_size_kb": 64, 00:08:00.510 "state": "configuring", 00:08:00.510 "raid_level": "raid0", 00:08:00.510 "superblock": false, 00:08:00.510 "num_base_bdevs": 3, 00:08:00.510 "num_base_bdevs_discovered": 2, 00:08:00.510 "num_base_bdevs_operational": 3, 00:08:00.510 "base_bdevs_list": [ 00:08:00.510 { 00:08:00.510 "name": null, 00:08:00.510 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:08:00.510 "is_configured": false, 00:08:00.510 "data_offset": 0, 00:08:00.510 "data_size": 65536 00:08:00.510 }, 00:08:00.510 { 00:08:00.510 "name": "BaseBdev2", 00:08:00.510 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:08:00.510 "is_configured": true, 00:08:00.510 "data_offset": 0, 00:08:00.510 "data_size": 65536 00:08:00.510 }, 00:08:00.510 { 00:08:00.510 "name": "BaseBdev3", 00:08:00.510 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:08:00.510 "is_configured": true, 00:08:00.510 "data_offset": 0, 00:08:00.510 "data_size": 65536 00:08:00.510 } 00:08:00.510 ] 00:08:00.510 }' 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.510 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c8c2af9f-8219-40c1-937f-a15d7258aead 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.079 [2024-09-30 14:07:05.679888] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:01.079 [2024-09-30 14:07:05.679988] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:01.079 [2024-09-30 14:07:05.679999] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:01.079 [2024-09-30 14:07:05.680254] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:08:01.079 [2024-09-30 14:07:05.680356] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:01.079 [2024-09-30 14:07:05.680370] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:08:01.079 [2024-09-30 14:07:05.680561] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:01.079 NewBaseBdev 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:01.079 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.080 [ 00:08:01.080 { 00:08:01.080 "name": "NewBaseBdev", 00:08:01.080 "aliases": [ 00:08:01.080 "c8c2af9f-8219-40c1-937f-a15d7258aead" 00:08:01.080 ], 00:08:01.080 "product_name": "Malloc disk", 00:08:01.080 "block_size": 512, 00:08:01.080 "num_blocks": 65536, 00:08:01.080 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:08:01.080 "assigned_rate_limits": { 00:08:01.080 "rw_ios_per_sec": 0, 00:08:01.080 "rw_mbytes_per_sec": 0, 00:08:01.080 "r_mbytes_per_sec": 0, 00:08:01.080 "w_mbytes_per_sec": 0 00:08:01.080 }, 00:08:01.080 "claimed": true, 00:08:01.080 "claim_type": "exclusive_write", 00:08:01.080 "zoned": false, 00:08:01.080 "supported_io_types": { 00:08:01.080 "read": true, 00:08:01.080 "write": true, 00:08:01.080 "unmap": true, 00:08:01.080 "flush": true, 00:08:01.080 "reset": true, 00:08:01.080 "nvme_admin": false, 00:08:01.080 "nvme_io": false, 00:08:01.080 "nvme_io_md": false, 00:08:01.080 "write_zeroes": true, 00:08:01.080 "zcopy": true, 00:08:01.080 "get_zone_info": false, 00:08:01.080 "zone_management": false, 00:08:01.080 "zone_append": false, 00:08:01.080 "compare": false, 00:08:01.080 "compare_and_write": false, 00:08:01.080 "abort": true, 00:08:01.080 "seek_hole": false, 00:08:01.080 "seek_data": false, 00:08:01.080 "copy": true, 00:08:01.080 "nvme_iov_md": false 00:08:01.080 }, 00:08:01.080 "memory_domains": [ 00:08:01.080 { 00:08:01.080 "dma_device_id": "system", 00:08:01.080 "dma_device_type": 1 00:08:01.080 }, 00:08:01.080 { 00:08:01.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.080 "dma_device_type": 2 00:08:01.080 } 00:08:01.080 ], 00:08:01.080 "driver_specific": {} 00:08:01.080 } 00:08:01.080 ] 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.080 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.339 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.339 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.339 "name": "Existed_Raid", 00:08:01.340 "uuid": "7f485c02-f473-4e79-a408-ed17dfb16dc2", 00:08:01.340 "strip_size_kb": 64, 00:08:01.340 "state": "online", 00:08:01.340 "raid_level": "raid0", 00:08:01.340 "superblock": false, 00:08:01.340 "num_base_bdevs": 3, 00:08:01.340 "num_base_bdevs_discovered": 3, 00:08:01.340 "num_base_bdevs_operational": 3, 00:08:01.340 "base_bdevs_list": [ 00:08:01.340 { 00:08:01.340 "name": "NewBaseBdev", 00:08:01.340 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:08:01.340 "is_configured": true, 00:08:01.340 "data_offset": 0, 00:08:01.340 "data_size": 65536 00:08:01.340 }, 00:08:01.340 { 00:08:01.340 "name": "BaseBdev2", 00:08:01.340 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:08:01.340 "is_configured": true, 00:08:01.340 "data_offset": 0, 00:08:01.340 "data_size": 65536 00:08:01.340 }, 00:08:01.340 { 00:08:01.340 "name": "BaseBdev3", 00:08:01.340 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:08:01.340 "is_configured": true, 00:08:01.340 "data_offset": 0, 00:08:01.340 "data_size": 65536 00:08:01.340 } 00:08:01.340 ] 00:08:01.340 }' 00:08:01.340 14:07:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.340 14:07:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.599 [2024-09-30 14:07:06.175339] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.599 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:01.599 "name": "Existed_Raid", 00:08:01.599 "aliases": [ 00:08:01.599 "7f485c02-f473-4e79-a408-ed17dfb16dc2" 00:08:01.599 ], 00:08:01.599 "product_name": "Raid Volume", 00:08:01.599 "block_size": 512, 00:08:01.599 "num_blocks": 196608, 00:08:01.599 "uuid": "7f485c02-f473-4e79-a408-ed17dfb16dc2", 00:08:01.599 "assigned_rate_limits": { 00:08:01.599 "rw_ios_per_sec": 0, 00:08:01.599 "rw_mbytes_per_sec": 0, 00:08:01.599 "r_mbytes_per_sec": 0, 00:08:01.599 "w_mbytes_per_sec": 0 00:08:01.599 }, 00:08:01.599 "claimed": false, 00:08:01.599 "zoned": false, 00:08:01.599 "supported_io_types": { 00:08:01.599 "read": true, 00:08:01.599 "write": true, 00:08:01.599 "unmap": true, 00:08:01.599 "flush": true, 00:08:01.599 "reset": true, 00:08:01.599 "nvme_admin": false, 00:08:01.600 "nvme_io": false, 00:08:01.600 "nvme_io_md": false, 00:08:01.600 "write_zeroes": true, 00:08:01.600 "zcopy": false, 00:08:01.600 "get_zone_info": false, 00:08:01.600 "zone_management": false, 00:08:01.600 "zone_append": false, 00:08:01.600 "compare": false, 00:08:01.600 "compare_and_write": false, 00:08:01.600 "abort": false, 00:08:01.600 "seek_hole": false, 00:08:01.600 "seek_data": false, 00:08:01.600 "copy": false, 00:08:01.600 "nvme_iov_md": false 00:08:01.600 }, 00:08:01.600 "memory_domains": [ 00:08:01.600 { 00:08:01.600 "dma_device_id": "system", 00:08:01.600 "dma_device_type": 1 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.600 "dma_device_type": 2 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "dma_device_id": "system", 00:08:01.600 "dma_device_type": 1 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.600 "dma_device_type": 2 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "dma_device_id": "system", 00:08:01.600 "dma_device_type": 1 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.600 "dma_device_type": 2 00:08:01.600 } 00:08:01.600 ], 00:08:01.600 "driver_specific": { 00:08:01.600 "raid": { 00:08:01.600 "uuid": "7f485c02-f473-4e79-a408-ed17dfb16dc2", 00:08:01.600 "strip_size_kb": 64, 00:08:01.600 "state": "online", 00:08:01.600 "raid_level": "raid0", 00:08:01.600 "superblock": false, 00:08:01.600 "num_base_bdevs": 3, 00:08:01.600 "num_base_bdevs_discovered": 3, 00:08:01.600 "num_base_bdevs_operational": 3, 00:08:01.600 "base_bdevs_list": [ 00:08:01.600 { 00:08:01.600 "name": "NewBaseBdev", 00:08:01.600 "uuid": "c8c2af9f-8219-40c1-937f-a15d7258aead", 00:08:01.600 "is_configured": true, 00:08:01.600 "data_offset": 0, 00:08:01.600 "data_size": 65536 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "name": "BaseBdev2", 00:08:01.600 "uuid": "47ab4145-b91a-4f94-871c-477931c4c7b3", 00:08:01.600 "is_configured": true, 00:08:01.600 "data_offset": 0, 00:08:01.600 "data_size": 65536 00:08:01.600 }, 00:08:01.600 { 00:08:01.600 "name": "BaseBdev3", 00:08:01.600 "uuid": "1959c96f-aeb7-4527-8cc6-46563210c6b0", 00:08:01.600 "is_configured": true, 00:08:01.600 "data_offset": 0, 00:08:01.600 "data_size": 65536 00:08:01.600 } 00:08:01.600 ] 00:08:01.600 } 00:08:01.600 } 00:08:01.600 }' 00:08:01.600 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:01.600 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:01.600 BaseBdev2 00:08:01.600 BaseBdev3' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.860 [2024-09-30 14:07:06.458596] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:01.860 [2024-09-30 14:07:06.458621] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:01.860 [2024-09-30 14:07:06.458684] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:01.860 [2024-09-30 14:07:06.458737] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:01.860 [2024-09-30 14:07:06.458746] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76081 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 76081 ']' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 76081 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76081 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76081' 00:08:01.860 killing process with pid 76081 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 76081 00:08:01.860 [2024-09-30 14:07:06.495825] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:01.860 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 76081 00:08:02.120 [2024-09-30 14:07:06.526344] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:02.120 14:07:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:02.120 00:08:02.120 real 0m8.565s 00:08:02.120 user 0m14.609s 00:08:02.120 sys 0m1.695s 00:08:02.120 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:02.120 ************************************ 00:08:02.120 END TEST raid_state_function_test 00:08:02.120 ************************************ 00:08:02.120 14:07:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.380 14:07:06 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:02.380 14:07:06 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:02.380 14:07:06 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:02.380 14:07:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:02.380 ************************************ 00:08:02.380 START TEST raid_state_function_test_sb 00:08:02.380 ************************************ 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 true 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:02.380 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=76680 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76680' 00:08:02.381 Process raid pid: 76680 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 76680 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 76680 ']' 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:02.381 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:02.381 14:07:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.381 [2024-09-30 14:07:06.934376] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:02.381 [2024-09-30 14:07:06.934569] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:02.641 [2024-09-30 14:07:07.065578] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:02.641 [2024-09-30 14:07:07.095110] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:02.641 [2024-09-30 14:07:07.140010] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:02.641 [2024-09-30 14:07:07.181009] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:02.641 [2024-09-30 14:07:07.181045] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.211 [2024-09-30 14:07:07.777699] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:03.211 [2024-09-30 14:07:07.777746] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:03.211 [2024-09-30 14:07:07.777759] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:03.211 [2024-09-30 14:07:07.777783] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:03.211 [2024-09-30 14:07:07.777794] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:03.211 [2024-09-30 14:07:07.777801] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.211 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.212 "name": "Existed_Raid", 00:08:03.212 "uuid": "9c54c14a-3625-4916-a051-1efc8c41ed99", 00:08:03.212 "strip_size_kb": 64, 00:08:03.212 "state": "configuring", 00:08:03.212 "raid_level": "raid0", 00:08:03.212 "superblock": true, 00:08:03.212 "num_base_bdevs": 3, 00:08:03.212 "num_base_bdevs_discovered": 0, 00:08:03.212 "num_base_bdevs_operational": 3, 00:08:03.212 "base_bdevs_list": [ 00:08:03.212 { 00:08:03.212 "name": "BaseBdev1", 00:08:03.212 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.212 "is_configured": false, 00:08:03.212 "data_offset": 0, 00:08:03.212 "data_size": 0 00:08:03.212 }, 00:08:03.212 { 00:08:03.212 "name": "BaseBdev2", 00:08:03.212 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.212 "is_configured": false, 00:08:03.212 "data_offset": 0, 00:08:03.212 "data_size": 0 00:08:03.212 }, 00:08:03.212 { 00:08:03.212 "name": "BaseBdev3", 00:08:03.212 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.212 "is_configured": false, 00:08:03.212 "data_offset": 0, 00:08:03.212 "data_size": 0 00:08:03.212 } 00:08:03.212 ] 00:08:03.212 }' 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.212 14:07:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.781 [2024-09-30 14:07:08.188894] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:03.781 [2024-09-30 14:07:08.188973] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:03.781 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.782 [2024-09-30 14:07:08.200884] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:03.782 [2024-09-30 14:07:08.200949] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:03.782 [2024-09-30 14:07:08.200996] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:03.782 [2024-09-30 14:07:08.201016] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:03.782 [2024-09-30 14:07:08.201043] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:03.782 [2024-09-30 14:07:08.201063] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.782 [2024-09-30 14:07:08.221458] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:03.782 BaseBdev1 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.782 [ 00:08:03.782 { 00:08:03.782 "name": "BaseBdev1", 00:08:03.782 "aliases": [ 00:08:03.782 "b912753a-cf2e-4902-bbe0-01d35f98975c" 00:08:03.782 ], 00:08:03.782 "product_name": "Malloc disk", 00:08:03.782 "block_size": 512, 00:08:03.782 "num_blocks": 65536, 00:08:03.782 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:03.782 "assigned_rate_limits": { 00:08:03.782 "rw_ios_per_sec": 0, 00:08:03.782 "rw_mbytes_per_sec": 0, 00:08:03.782 "r_mbytes_per_sec": 0, 00:08:03.782 "w_mbytes_per_sec": 0 00:08:03.782 }, 00:08:03.782 "claimed": true, 00:08:03.782 "claim_type": "exclusive_write", 00:08:03.782 "zoned": false, 00:08:03.782 "supported_io_types": { 00:08:03.782 "read": true, 00:08:03.782 "write": true, 00:08:03.782 "unmap": true, 00:08:03.782 "flush": true, 00:08:03.782 "reset": true, 00:08:03.782 "nvme_admin": false, 00:08:03.782 "nvme_io": false, 00:08:03.782 "nvme_io_md": false, 00:08:03.782 "write_zeroes": true, 00:08:03.782 "zcopy": true, 00:08:03.782 "get_zone_info": false, 00:08:03.782 "zone_management": false, 00:08:03.782 "zone_append": false, 00:08:03.782 "compare": false, 00:08:03.782 "compare_and_write": false, 00:08:03.782 "abort": true, 00:08:03.782 "seek_hole": false, 00:08:03.782 "seek_data": false, 00:08:03.782 "copy": true, 00:08:03.782 "nvme_iov_md": false 00:08:03.782 }, 00:08:03.782 "memory_domains": [ 00:08:03.782 { 00:08:03.782 "dma_device_id": "system", 00:08:03.782 "dma_device_type": 1 00:08:03.782 }, 00:08:03.782 { 00:08:03.782 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.782 "dma_device_type": 2 00:08:03.782 } 00:08:03.782 ], 00:08:03.782 "driver_specific": {} 00:08:03.782 } 00:08:03.782 ] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.782 "name": "Existed_Raid", 00:08:03.782 "uuid": "ef9ff759-8165-4f6d-b3f1-971bb76f2928", 00:08:03.782 "strip_size_kb": 64, 00:08:03.782 "state": "configuring", 00:08:03.782 "raid_level": "raid0", 00:08:03.782 "superblock": true, 00:08:03.782 "num_base_bdevs": 3, 00:08:03.782 "num_base_bdevs_discovered": 1, 00:08:03.782 "num_base_bdevs_operational": 3, 00:08:03.782 "base_bdevs_list": [ 00:08:03.782 { 00:08:03.782 "name": "BaseBdev1", 00:08:03.782 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:03.782 "is_configured": true, 00:08:03.782 "data_offset": 2048, 00:08:03.782 "data_size": 63488 00:08:03.782 }, 00:08:03.782 { 00:08:03.782 "name": "BaseBdev2", 00:08:03.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.782 "is_configured": false, 00:08:03.782 "data_offset": 0, 00:08:03.782 "data_size": 0 00:08:03.782 }, 00:08:03.782 { 00:08:03.782 "name": "BaseBdev3", 00:08:03.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.782 "is_configured": false, 00:08:03.782 "data_offset": 0, 00:08:03.782 "data_size": 0 00:08:03.782 } 00:08:03.782 ] 00:08:03.782 }' 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.782 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.350 [2024-09-30 14:07:08.708651] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:04.350 [2024-09-30 14:07:08.708698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.350 [2024-09-30 14:07:08.720675] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:04.350 [2024-09-30 14:07:08.722543] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:04.350 [2024-09-30 14:07:08.722599] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:04.350 [2024-09-30 14:07:08.722612] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:04.350 [2024-09-30 14:07:08.722620] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.350 "name": "Existed_Raid", 00:08:04.350 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:04.350 "strip_size_kb": 64, 00:08:04.350 "state": "configuring", 00:08:04.350 "raid_level": "raid0", 00:08:04.350 "superblock": true, 00:08:04.350 "num_base_bdevs": 3, 00:08:04.350 "num_base_bdevs_discovered": 1, 00:08:04.350 "num_base_bdevs_operational": 3, 00:08:04.350 "base_bdevs_list": [ 00:08:04.350 { 00:08:04.350 "name": "BaseBdev1", 00:08:04.350 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:04.350 "is_configured": true, 00:08:04.350 "data_offset": 2048, 00:08:04.350 "data_size": 63488 00:08:04.350 }, 00:08:04.350 { 00:08:04.350 "name": "BaseBdev2", 00:08:04.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:04.350 "is_configured": false, 00:08:04.350 "data_offset": 0, 00:08:04.350 "data_size": 0 00:08:04.350 }, 00:08:04.350 { 00:08:04.350 "name": "BaseBdev3", 00:08:04.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:04.350 "is_configured": false, 00:08:04.350 "data_offset": 0, 00:08:04.350 "data_size": 0 00:08:04.350 } 00:08:04.350 ] 00:08:04.350 }' 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.350 14:07:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.609 [2024-09-30 14:07:09.168186] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:04.609 BaseBdev2 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.609 [ 00:08:04.609 { 00:08:04.609 "name": "BaseBdev2", 00:08:04.609 "aliases": [ 00:08:04.609 "6bcbf207-67e4-4511-8301-84c20ae824b4" 00:08:04.609 ], 00:08:04.609 "product_name": "Malloc disk", 00:08:04.609 "block_size": 512, 00:08:04.609 "num_blocks": 65536, 00:08:04.609 "uuid": "6bcbf207-67e4-4511-8301-84c20ae824b4", 00:08:04.609 "assigned_rate_limits": { 00:08:04.609 "rw_ios_per_sec": 0, 00:08:04.609 "rw_mbytes_per_sec": 0, 00:08:04.609 "r_mbytes_per_sec": 0, 00:08:04.609 "w_mbytes_per_sec": 0 00:08:04.609 }, 00:08:04.609 "claimed": true, 00:08:04.609 "claim_type": "exclusive_write", 00:08:04.609 "zoned": false, 00:08:04.609 "supported_io_types": { 00:08:04.609 "read": true, 00:08:04.609 "write": true, 00:08:04.609 "unmap": true, 00:08:04.609 "flush": true, 00:08:04.609 "reset": true, 00:08:04.609 "nvme_admin": false, 00:08:04.609 "nvme_io": false, 00:08:04.609 "nvme_io_md": false, 00:08:04.609 "write_zeroes": true, 00:08:04.609 "zcopy": true, 00:08:04.609 "get_zone_info": false, 00:08:04.609 "zone_management": false, 00:08:04.609 "zone_append": false, 00:08:04.609 "compare": false, 00:08:04.609 "compare_and_write": false, 00:08:04.609 "abort": true, 00:08:04.609 "seek_hole": false, 00:08:04.609 "seek_data": false, 00:08:04.609 "copy": true, 00:08:04.609 "nvme_iov_md": false 00:08:04.609 }, 00:08:04.609 "memory_domains": [ 00:08:04.609 { 00:08:04.609 "dma_device_id": "system", 00:08:04.609 "dma_device_type": 1 00:08:04.609 }, 00:08:04.609 { 00:08:04.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.609 "dma_device_type": 2 00:08:04.609 } 00:08:04.609 ], 00:08:04.609 "driver_specific": {} 00:08:04.609 } 00:08:04.609 ] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.609 "name": "Existed_Raid", 00:08:04.609 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:04.609 "strip_size_kb": 64, 00:08:04.609 "state": "configuring", 00:08:04.609 "raid_level": "raid0", 00:08:04.609 "superblock": true, 00:08:04.609 "num_base_bdevs": 3, 00:08:04.609 "num_base_bdevs_discovered": 2, 00:08:04.609 "num_base_bdevs_operational": 3, 00:08:04.609 "base_bdevs_list": [ 00:08:04.609 { 00:08:04.609 "name": "BaseBdev1", 00:08:04.609 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:04.609 "is_configured": true, 00:08:04.609 "data_offset": 2048, 00:08:04.609 "data_size": 63488 00:08:04.609 }, 00:08:04.609 { 00:08:04.609 "name": "BaseBdev2", 00:08:04.609 "uuid": "6bcbf207-67e4-4511-8301-84c20ae824b4", 00:08:04.609 "is_configured": true, 00:08:04.609 "data_offset": 2048, 00:08:04.609 "data_size": 63488 00:08:04.609 }, 00:08:04.609 { 00:08:04.609 "name": "BaseBdev3", 00:08:04.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:04.609 "is_configured": false, 00:08:04.609 "data_offset": 0, 00:08:04.609 "data_size": 0 00:08:04.609 } 00:08:04.609 ] 00:08:04.609 }' 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.609 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.177 [2024-09-30 14:07:09.654027] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:05.177 BaseBdev3 00:08:05.177 [2024-09-30 14:07:09.654288] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:05.177 [2024-09-30 14:07:09.654305] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:05.177 [2024-09-30 14:07:09.654605] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:05.177 [2024-09-30 14:07:09.654748] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:05.177 [2024-09-30 14:07:09.654761] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:05.177 [2024-09-30 14:07:09.654883] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.177 [ 00:08:05.177 { 00:08:05.177 "name": "BaseBdev3", 00:08:05.177 "aliases": [ 00:08:05.177 "68d2d6e9-85cc-44a8-b110-764287c63764" 00:08:05.177 ], 00:08:05.177 "product_name": "Malloc disk", 00:08:05.177 "block_size": 512, 00:08:05.177 "num_blocks": 65536, 00:08:05.177 "uuid": "68d2d6e9-85cc-44a8-b110-764287c63764", 00:08:05.177 "assigned_rate_limits": { 00:08:05.177 "rw_ios_per_sec": 0, 00:08:05.177 "rw_mbytes_per_sec": 0, 00:08:05.177 "r_mbytes_per_sec": 0, 00:08:05.177 "w_mbytes_per_sec": 0 00:08:05.177 }, 00:08:05.177 "claimed": true, 00:08:05.177 "claim_type": "exclusive_write", 00:08:05.177 "zoned": false, 00:08:05.177 "supported_io_types": { 00:08:05.177 "read": true, 00:08:05.177 "write": true, 00:08:05.177 "unmap": true, 00:08:05.177 "flush": true, 00:08:05.177 "reset": true, 00:08:05.177 "nvme_admin": false, 00:08:05.177 "nvme_io": false, 00:08:05.177 "nvme_io_md": false, 00:08:05.177 "write_zeroes": true, 00:08:05.177 "zcopy": true, 00:08:05.177 "get_zone_info": false, 00:08:05.177 "zone_management": false, 00:08:05.177 "zone_append": false, 00:08:05.177 "compare": false, 00:08:05.177 "compare_and_write": false, 00:08:05.177 "abort": true, 00:08:05.177 "seek_hole": false, 00:08:05.177 "seek_data": false, 00:08:05.177 "copy": true, 00:08:05.177 "nvme_iov_md": false 00:08:05.177 }, 00:08:05.177 "memory_domains": [ 00:08:05.177 { 00:08:05.177 "dma_device_id": "system", 00:08:05.177 "dma_device_type": 1 00:08:05.177 }, 00:08:05.177 { 00:08:05.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.177 "dma_device_type": 2 00:08:05.177 } 00:08:05.177 ], 00:08:05.177 "driver_specific": {} 00:08:05.177 } 00:08:05.177 ] 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:05.177 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:05.178 "name": "Existed_Raid", 00:08:05.178 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:05.178 "strip_size_kb": 64, 00:08:05.178 "state": "online", 00:08:05.178 "raid_level": "raid0", 00:08:05.178 "superblock": true, 00:08:05.178 "num_base_bdevs": 3, 00:08:05.178 "num_base_bdevs_discovered": 3, 00:08:05.178 "num_base_bdevs_operational": 3, 00:08:05.178 "base_bdevs_list": [ 00:08:05.178 { 00:08:05.178 "name": "BaseBdev1", 00:08:05.178 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:05.178 "is_configured": true, 00:08:05.178 "data_offset": 2048, 00:08:05.178 "data_size": 63488 00:08:05.178 }, 00:08:05.178 { 00:08:05.178 "name": "BaseBdev2", 00:08:05.178 "uuid": "6bcbf207-67e4-4511-8301-84c20ae824b4", 00:08:05.178 "is_configured": true, 00:08:05.178 "data_offset": 2048, 00:08:05.178 "data_size": 63488 00:08:05.178 }, 00:08:05.178 { 00:08:05.178 "name": "BaseBdev3", 00:08:05.178 "uuid": "68d2d6e9-85cc-44a8-b110-764287c63764", 00:08:05.178 "is_configured": true, 00:08:05.178 "data_offset": 2048, 00:08:05.178 "data_size": 63488 00:08:05.178 } 00:08:05.178 ] 00:08:05.178 }' 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:05.178 14:07:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:05.745 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.746 [2024-09-30 14:07:10.165507] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:05.746 "name": "Existed_Raid", 00:08:05.746 "aliases": [ 00:08:05.746 "494f47fa-7798-4280-8a45-2cc9976e3759" 00:08:05.746 ], 00:08:05.746 "product_name": "Raid Volume", 00:08:05.746 "block_size": 512, 00:08:05.746 "num_blocks": 190464, 00:08:05.746 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:05.746 "assigned_rate_limits": { 00:08:05.746 "rw_ios_per_sec": 0, 00:08:05.746 "rw_mbytes_per_sec": 0, 00:08:05.746 "r_mbytes_per_sec": 0, 00:08:05.746 "w_mbytes_per_sec": 0 00:08:05.746 }, 00:08:05.746 "claimed": false, 00:08:05.746 "zoned": false, 00:08:05.746 "supported_io_types": { 00:08:05.746 "read": true, 00:08:05.746 "write": true, 00:08:05.746 "unmap": true, 00:08:05.746 "flush": true, 00:08:05.746 "reset": true, 00:08:05.746 "nvme_admin": false, 00:08:05.746 "nvme_io": false, 00:08:05.746 "nvme_io_md": false, 00:08:05.746 "write_zeroes": true, 00:08:05.746 "zcopy": false, 00:08:05.746 "get_zone_info": false, 00:08:05.746 "zone_management": false, 00:08:05.746 "zone_append": false, 00:08:05.746 "compare": false, 00:08:05.746 "compare_and_write": false, 00:08:05.746 "abort": false, 00:08:05.746 "seek_hole": false, 00:08:05.746 "seek_data": false, 00:08:05.746 "copy": false, 00:08:05.746 "nvme_iov_md": false 00:08:05.746 }, 00:08:05.746 "memory_domains": [ 00:08:05.746 { 00:08:05.746 "dma_device_id": "system", 00:08:05.746 "dma_device_type": 1 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.746 "dma_device_type": 2 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "dma_device_id": "system", 00:08:05.746 "dma_device_type": 1 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.746 "dma_device_type": 2 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "dma_device_id": "system", 00:08:05.746 "dma_device_type": 1 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.746 "dma_device_type": 2 00:08:05.746 } 00:08:05.746 ], 00:08:05.746 "driver_specific": { 00:08:05.746 "raid": { 00:08:05.746 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:05.746 "strip_size_kb": 64, 00:08:05.746 "state": "online", 00:08:05.746 "raid_level": "raid0", 00:08:05.746 "superblock": true, 00:08:05.746 "num_base_bdevs": 3, 00:08:05.746 "num_base_bdevs_discovered": 3, 00:08:05.746 "num_base_bdevs_operational": 3, 00:08:05.746 "base_bdevs_list": [ 00:08:05.746 { 00:08:05.746 "name": "BaseBdev1", 00:08:05.746 "uuid": "b912753a-cf2e-4902-bbe0-01d35f98975c", 00:08:05.746 "is_configured": true, 00:08:05.746 "data_offset": 2048, 00:08:05.746 "data_size": 63488 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "name": "BaseBdev2", 00:08:05.746 "uuid": "6bcbf207-67e4-4511-8301-84c20ae824b4", 00:08:05.746 "is_configured": true, 00:08:05.746 "data_offset": 2048, 00:08:05.746 "data_size": 63488 00:08:05.746 }, 00:08:05.746 { 00:08:05.746 "name": "BaseBdev3", 00:08:05.746 "uuid": "68d2d6e9-85cc-44a8-b110-764287c63764", 00:08:05.746 "is_configured": true, 00:08:05.746 "data_offset": 2048, 00:08:05.746 "data_size": 63488 00:08:05.746 } 00:08:05.746 ] 00:08:05.746 } 00:08:05.746 } 00:08:05.746 }' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:05.746 BaseBdev2 00:08:05.746 BaseBdev3' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.746 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.005 [2024-09-30 14:07:10.416794] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:06.005 [2024-09-30 14:07:10.416820] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:06.005 [2024-09-30 14:07:10.416879] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.005 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.005 "name": "Existed_Raid", 00:08:06.005 "uuid": "494f47fa-7798-4280-8a45-2cc9976e3759", 00:08:06.005 "strip_size_kb": 64, 00:08:06.005 "state": "offline", 00:08:06.005 "raid_level": "raid0", 00:08:06.005 "superblock": true, 00:08:06.005 "num_base_bdevs": 3, 00:08:06.005 "num_base_bdevs_discovered": 2, 00:08:06.005 "num_base_bdevs_operational": 2, 00:08:06.005 "base_bdevs_list": [ 00:08:06.005 { 00:08:06.005 "name": null, 00:08:06.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.005 "is_configured": false, 00:08:06.005 "data_offset": 0, 00:08:06.005 "data_size": 63488 00:08:06.005 }, 00:08:06.005 { 00:08:06.005 "name": "BaseBdev2", 00:08:06.005 "uuid": "6bcbf207-67e4-4511-8301-84c20ae824b4", 00:08:06.005 "is_configured": true, 00:08:06.005 "data_offset": 2048, 00:08:06.005 "data_size": 63488 00:08:06.005 }, 00:08:06.005 { 00:08:06.005 "name": "BaseBdev3", 00:08:06.005 "uuid": "68d2d6e9-85cc-44a8-b110-764287c63764", 00:08:06.005 "is_configured": true, 00:08:06.006 "data_offset": 2048, 00:08:06.006 "data_size": 63488 00:08:06.006 } 00:08:06.006 ] 00:08:06.006 }' 00:08:06.006 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.006 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.264 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.265 [2024-09-30 14:07:10.847172] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.265 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.265 [2024-09-30 14:07:10.909935] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:06.265 [2024-09-30 14:07:10.910028] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 BaseBdev2 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 14:07:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 [ 00:08:06.524 { 00:08:06.524 "name": "BaseBdev2", 00:08:06.524 "aliases": [ 00:08:06.524 "2730fcef-132e-4aa4-a51b-df351df08629" 00:08:06.524 ], 00:08:06.524 "product_name": "Malloc disk", 00:08:06.524 "block_size": 512, 00:08:06.524 "num_blocks": 65536, 00:08:06.524 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:06.524 "assigned_rate_limits": { 00:08:06.524 "rw_ios_per_sec": 0, 00:08:06.524 "rw_mbytes_per_sec": 0, 00:08:06.524 "r_mbytes_per_sec": 0, 00:08:06.524 "w_mbytes_per_sec": 0 00:08:06.524 }, 00:08:06.524 "claimed": false, 00:08:06.524 "zoned": false, 00:08:06.524 "supported_io_types": { 00:08:06.524 "read": true, 00:08:06.524 "write": true, 00:08:06.524 "unmap": true, 00:08:06.524 "flush": true, 00:08:06.524 "reset": true, 00:08:06.524 "nvme_admin": false, 00:08:06.524 "nvme_io": false, 00:08:06.524 "nvme_io_md": false, 00:08:06.524 "write_zeroes": true, 00:08:06.524 "zcopy": true, 00:08:06.524 "get_zone_info": false, 00:08:06.524 "zone_management": false, 00:08:06.524 "zone_append": false, 00:08:06.524 "compare": false, 00:08:06.524 "compare_and_write": false, 00:08:06.524 "abort": true, 00:08:06.524 "seek_hole": false, 00:08:06.524 "seek_data": false, 00:08:06.524 "copy": true, 00:08:06.524 "nvme_iov_md": false 00:08:06.524 }, 00:08:06.524 "memory_domains": [ 00:08:06.524 { 00:08:06.524 "dma_device_id": "system", 00:08:06.524 "dma_device_type": 1 00:08:06.524 }, 00:08:06.524 { 00:08:06.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.524 "dma_device_type": 2 00:08:06.524 } 00:08:06.524 ], 00:08:06.524 "driver_specific": {} 00:08:06.524 } 00:08:06.524 ] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 BaseBdev3 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.524 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.524 [ 00:08:06.524 { 00:08:06.524 "name": "BaseBdev3", 00:08:06.524 "aliases": [ 00:08:06.524 "5c8647c6-87dd-484f-9a76-bb6d6fa3a985" 00:08:06.524 ], 00:08:06.524 "product_name": "Malloc disk", 00:08:06.524 "block_size": 512, 00:08:06.524 "num_blocks": 65536, 00:08:06.524 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:06.524 "assigned_rate_limits": { 00:08:06.524 "rw_ios_per_sec": 0, 00:08:06.524 "rw_mbytes_per_sec": 0, 00:08:06.524 "r_mbytes_per_sec": 0, 00:08:06.524 "w_mbytes_per_sec": 0 00:08:06.524 }, 00:08:06.524 "claimed": false, 00:08:06.524 "zoned": false, 00:08:06.525 "supported_io_types": { 00:08:06.525 "read": true, 00:08:06.525 "write": true, 00:08:06.525 "unmap": true, 00:08:06.525 "flush": true, 00:08:06.525 "reset": true, 00:08:06.525 "nvme_admin": false, 00:08:06.525 "nvme_io": false, 00:08:06.525 "nvme_io_md": false, 00:08:06.525 "write_zeroes": true, 00:08:06.525 "zcopy": true, 00:08:06.525 "get_zone_info": false, 00:08:06.525 "zone_management": false, 00:08:06.525 "zone_append": false, 00:08:06.525 "compare": false, 00:08:06.525 "compare_and_write": false, 00:08:06.525 "abort": true, 00:08:06.525 "seek_hole": false, 00:08:06.525 "seek_data": false, 00:08:06.525 "copy": true, 00:08:06.525 "nvme_iov_md": false 00:08:06.525 }, 00:08:06.525 "memory_domains": [ 00:08:06.525 { 00:08:06.525 "dma_device_id": "system", 00:08:06.525 "dma_device_type": 1 00:08:06.525 }, 00:08:06.525 { 00:08:06.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.525 "dma_device_type": 2 00:08:06.525 } 00:08:06.525 ], 00:08:06.525 "driver_specific": {} 00:08:06.525 } 00:08:06.525 ] 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.525 [2024-09-30 14:07:11.084188] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:06.525 [2024-09-30 14:07:11.084271] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:06.525 [2024-09-30 14:07:11.084293] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:06.525 [2024-09-30 14:07:11.086011] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.525 "name": "Existed_Raid", 00:08:06.525 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:06.525 "strip_size_kb": 64, 00:08:06.525 "state": "configuring", 00:08:06.525 "raid_level": "raid0", 00:08:06.525 "superblock": true, 00:08:06.525 "num_base_bdevs": 3, 00:08:06.525 "num_base_bdevs_discovered": 2, 00:08:06.525 "num_base_bdevs_operational": 3, 00:08:06.525 "base_bdevs_list": [ 00:08:06.525 { 00:08:06.525 "name": "BaseBdev1", 00:08:06.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.525 "is_configured": false, 00:08:06.525 "data_offset": 0, 00:08:06.525 "data_size": 0 00:08:06.525 }, 00:08:06.525 { 00:08:06.525 "name": "BaseBdev2", 00:08:06.525 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:06.525 "is_configured": true, 00:08:06.525 "data_offset": 2048, 00:08:06.525 "data_size": 63488 00:08:06.525 }, 00:08:06.525 { 00:08:06.525 "name": "BaseBdev3", 00:08:06.525 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:06.525 "is_configured": true, 00:08:06.525 "data_offset": 2048, 00:08:06.525 "data_size": 63488 00:08:06.525 } 00:08:06.525 ] 00:08:06.525 }' 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.525 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.093 [2024-09-30 14:07:11.495511] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.093 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.094 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.094 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.094 "name": "Existed_Raid", 00:08:07.094 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:07.094 "strip_size_kb": 64, 00:08:07.094 "state": "configuring", 00:08:07.094 "raid_level": "raid0", 00:08:07.094 "superblock": true, 00:08:07.094 "num_base_bdevs": 3, 00:08:07.094 "num_base_bdevs_discovered": 1, 00:08:07.094 "num_base_bdevs_operational": 3, 00:08:07.094 "base_bdevs_list": [ 00:08:07.094 { 00:08:07.094 "name": "BaseBdev1", 00:08:07.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:07.094 "is_configured": false, 00:08:07.094 "data_offset": 0, 00:08:07.094 "data_size": 0 00:08:07.094 }, 00:08:07.094 { 00:08:07.094 "name": null, 00:08:07.094 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:07.094 "is_configured": false, 00:08:07.094 "data_offset": 0, 00:08:07.094 "data_size": 63488 00:08:07.094 }, 00:08:07.094 { 00:08:07.094 "name": "BaseBdev3", 00:08:07.094 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:07.094 "is_configured": true, 00:08:07.094 "data_offset": 2048, 00:08:07.094 "data_size": 63488 00:08:07.094 } 00:08:07.094 ] 00:08:07.094 }' 00:08:07.094 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.094 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.353 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.353 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.353 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.353 14:07:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:07.353 14:07:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.612 [2024-09-30 14:07:12.021404] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:07.612 BaseBdev1 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.612 [ 00:08:07.612 { 00:08:07.612 "name": "BaseBdev1", 00:08:07.612 "aliases": [ 00:08:07.612 "6db25d93-61a3-4c6c-9049-c38bc8a63297" 00:08:07.612 ], 00:08:07.612 "product_name": "Malloc disk", 00:08:07.612 "block_size": 512, 00:08:07.612 "num_blocks": 65536, 00:08:07.612 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:07.612 "assigned_rate_limits": { 00:08:07.612 "rw_ios_per_sec": 0, 00:08:07.612 "rw_mbytes_per_sec": 0, 00:08:07.612 "r_mbytes_per_sec": 0, 00:08:07.612 "w_mbytes_per_sec": 0 00:08:07.612 }, 00:08:07.612 "claimed": true, 00:08:07.612 "claim_type": "exclusive_write", 00:08:07.612 "zoned": false, 00:08:07.612 "supported_io_types": { 00:08:07.612 "read": true, 00:08:07.612 "write": true, 00:08:07.612 "unmap": true, 00:08:07.612 "flush": true, 00:08:07.612 "reset": true, 00:08:07.612 "nvme_admin": false, 00:08:07.612 "nvme_io": false, 00:08:07.612 "nvme_io_md": false, 00:08:07.612 "write_zeroes": true, 00:08:07.612 "zcopy": true, 00:08:07.612 "get_zone_info": false, 00:08:07.612 "zone_management": false, 00:08:07.612 "zone_append": false, 00:08:07.612 "compare": false, 00:08:07.612 "compare_and_write": false, 00:08:07.612 "abort": true, 00:08:07.612 "seek_hole": false, 00:08:07.612 "seek_data": false, 00:08:07.612 "copy": true, 00:08:07.612 "nvme_iov_md": false 00:08:07.612 }, 00:08:07.612 "memory_domains": [ 00:08:07.612 { 00:08:07.612 "dma_device_id": "system", 00:08:07.612 "dma_device_type": 1 00:08:07.612 }, 00:08:07.612 { 00:08:07.612 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:07.612 "dma_device_type": 2 00:08:07.612 } 00:08:07.612 ], 00:08:07.612 "driver_specific": {} 00:08:07.612 } 00:08:07.612 ] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.612 "name": "Existed_Raid", 00:08:07.612 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:07.612 "strip_size_kb": 64, 00:08:07.612 "state": "configuring", 00:08:07.612 "raid_level": "raid0", 00:08:07.612 "superblock": true, 00:08:07.612 "num_base_bdevs": 3, 00:08:07.612 "num_base_bdevs_discovered": 2, 00:08:07.612 "num_base_bdevs_operational": 3, 00:08:07.612 "base_bdevs_list": [ 00:08:07.612 { 00:08:07.612 "name": "BaseBdev1", 00:08:07.612 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:07.612 "is_configured": true, 00:08:07.612 "data_offset": 2048, 00:08:07.612 "data_size": 63488 00:08:07.612 }, 00:08:07.612 { 00:08:07.612 "name": null, 00:08:07.612 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:07.612 "is_configured": false, 00:08:07.612 "data_offset": 0, 00:08:07.612 "data_size": 63488 00:08:07.612 }, 00:08:07.612 { 00:08:07.612 "name": "BaseBdev3", 00:08:07.612 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:07.612 "is_configured": true, 00:08:07.612 "data_offset": 2048, 00:08:07.612 "data_size": 63488 00:08:07.612 } 00:08:07.612 ] 00:08:07.612 }' 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.612 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.871 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.871 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:07.871 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.871 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.130 [2024-09-30 14:07:12.560557] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.130 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.130 "name": "Existed_Raid", 00:08:08.130 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:08.130 "strip_size_kb": 64, 00:08:08.130 "state": "configuring", 00:08:08.130 "raid_level": "raid0", 00:08:08.130 "superblock": true, 00:08:08.130 "num_base_bdevs": 3, 00:08:08.130 "num_base_bdevs_discovered": 1, 00:08:08.130 "num_base_bdevs_operational": 3, 00:08:08.130 "base_bdevs_list": [ 00:08:08.130 { 00:08:08.130 "name": "BaseBdev1", 00:08:08.131 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:08.131 "is_configured": true, 00:08:08.131 "data_offset": 2048, 00:08:08.131 "data_size": 63488 00:08:08.131 }, 00:08:08.131 { 00:08:08.131 "name": null, 00:08:08.131 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:08.131 "is_configured": false, 00:08:08.131 "data_offset": 0, 00:08:08.131 "data_size": 63488 00:08:08.131 }, 00:08:08.131 { 00:08:08.131 "name": null, 00:08:08.131 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:08.131 "is_configured": false, 00:08:08.131 "data_offset": 0, 00:08:08.131 "data_size": 63488 00:08:08.131 } 00:08:08.131 ] 00:08:08.131 }' 00:08:08.131 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.131 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.390 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:08.390 14:07:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.390 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.390 14:07:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.390 [2024-09-30 14:07:13.019773] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.390 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.650 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.650 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.650 "name": "Existed_Raid", 00:08:08.650 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:08.650 "strip_size_kb": 64, 00:08:08.650 "state": "configuring", 00:08:08.650 "raid_level": "raid0", 00:08:08.650 "superblock": true, 00:08:08.650 "num_base_bdevs": 3, 00:08:08.650 "num_base_bdevs_discovered": 2, 00:08:08.650 "num_base_bdevs_operational": 3, 00:08:08.650 "base_bdevs_list": [ 00:08:08.650 { 00:08:08.650 "name": "BaseBdev1", 00:08:08.650 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:08.650 "is_configured": true, 00:08:08.650 "data_offset": 2048, 00:08:08.650 "data_size": 63488 00:08:08.650 }, 00:08:08.650 { 00:08:08.650 "name": null, 00:08:08.650 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:08.650 "is_configured": false, 00:08:08.650 "data_offset": 0, 00:08:08.650 "data_size": 63488 00:08:08.650 }, 00:08:08.650 { 00:08:08.650 "name": "BaseBdev3", 00:08:08.650 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:08.650 "is_configured": true, 00:08:08.650 "data_offset": 2048, 00:08:08.650 "data_size": 63488 00:08:08.650 } 00:08:08.650 ] 00:08:08.650 }' 00:08:08.650 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.650 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.910 [2024-09-30 14:07:13.443111] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.910 "name": "Existed_Raid", 00:08:08.910 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:08.910 "strip_size_kb": 64, 00:08:08.910 "state": "configuring", 00:08:08.910 "raid_level": "raid0", 00:08:08.910 "superblock": true, 00:08:08.910 "num_base_bdevs": 3, 00:08:08.910 "num_base_bdevs_discovered": 1, 00:08:08.910 "num_base_bdevs_operational": 3, 00:08:08.910 "base_bdevs_list": [ 00:08:08.910 { 00:08:08.910 "name": null, 00:08:08.910 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:08.910 "is_configured": false, 00:08:08.910 "data_offset": 0, 00:08:08.910 "data_size": 63488 00:08:08.910 }, 00:08:08.910 { 00:08:08.910 "name": null, 00:08:08.910 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:08.910 "is_configured": false, 00:08:08.910 "data_offset": 0, 00:08:08.910 "data_size": 63488 00:08:08.910 }, 00:08:08.910 { 00:08:08.910 "name": "BaseBdev3", 00:08:08.910 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:08.910 "is_configured": true, 00:08:08.910 "data_offset": 2048, 00:08:08.910 "data_size": 63488 00:08:08.910 } 00:08:08.910 ] 00:08:08.910 }' 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.910 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.479 [2024-09-30 14:07:13.908725] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:09.479 "name": "Existed_Raid", 00:08:09.479 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:09.479 "strip_size_kb": 64, 00:08:09.479 "state": "configuring", 00:08:09.479 "raid_level": "raid0", 00:08:09.479 "superblock": true, 00:08:09.479 "num_base_bdevs": 3, 00:08:09.479 "num_base_bdevs_discovered": 2, 00:08:09.479 "num_base_bdevs_operational": 3, 00:08:09.479 "base_bdevs_list": [ 00:08:09.479 { 00:08:09.479 "name": null, 00:08:09.479 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:09.479 "is_configured": false, 00:08:09.479 "data_offset": 0, 00:08:09.479 "data_size": 63488 00:08:09.479 }, 00:08:09.479 { 00:08:09.479 "name": "BaseBdev2", 00:08:09.479 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:09.479 "is_configured": true, 00:08:09.479 "data_offset": 2048, 00:08:09.479 "data_size": 63488 00:08:09.479 }, 00:08:09.479 { 00:08:09.479 "name": "BaseBdev3", 00:08:09.479 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:09.479 "is_configured": true, 00:08:09.479 "data_offset": 2048, 00:08:09.479 "data_size": 63488 00:08:09.479 } 00:08:09.479 ] 00:08:09.479 }' 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:09.479 14:07:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.739 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:09.739 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.739 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.739 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.739 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6db25d93-61a3-4c6c-9049-c38bc8a63297 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.999 [2024-09-30 14:07:14.454659] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:09.999 [2024-09-30 14:07:14.454812] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:09.999 [2024-09-30 14:07:14.454825] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:09.999 [2024-09-30 14:07:14.455073] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:08:09.999 [2024-09-30 14:07:14.455174] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:09.999 [2024-09-30 14:07:14.455188] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:08:09.999 NewBaseBdev 00:08:09.999 [2024-09-30 14:07:14.455280] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.999 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.999 [ 00:08:09.999 { 00:08:09.999 "name": "NewBaseBdev", 00:08:09.999 "aliases": [ 00:08:10.000 "6db25d93-61a3-4c6c-9049-c38bc8a63297" 00:08:10.000 ], 00:08:10.000 "product_name": "Malloc disk", 00:08:10.000 "block_size": 512, 00:08:10.000 "num_blocks": 65536, 00:08:10.000 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:10.000 "assigned_rate_limits": { 00:08:10.000 "rw_ios_per_sec": 0, 00:08:10.000 "rw_mbytes_per_sec": 0, 00:08:10.000 "r_mbytes_per_sec": 0, 00:08:10.000 "w_mbytes_per_sec": 0 00:08:10.000 }, 00:08:10.000 "claimed": true, 00:08:10.000 "claim_type": "exclusive_write", 00:08:10.000 "zoned": false, 00:08:10.000 "supported_io_types": { 00:08:10.000 "read": true, 00:08:10.000 "write": true, 00:08:10.000 "unmap": true, 00:08:10.000 "flush": true, 00:08:10.000 "reset": true, 00:08:10.000 "nvme_admin": false, 00:08:10.000 "nvme_io": false, 00:08:10.000 "nvme_io_md": false, 00:08:10.000 "write_zeroes": true, 00:08:10.000 "zcopy": true, 00:08:10.000 "get_zone_info": false, 00:08:10.000 "zone_management": false, 00:08:10.000 "zone_append": false, 00:08:10.000 "compare": false, 00:08:10.000 "compare_and_write": false, 00:08:10.000 "abort": true, 00:08:10.000 "seek_hole": false, 00:08:10.000 "seek_data": false, 00:08:10.000 "copy": true, 00:08:10.000 "nvme_iov_md": false 00:08:10.000 }, 00:08:10.000 "memory_domains": [ 00:08:10.000 { 00:08:10.000 "dma_device_id": "system", 00:08:10.000 "dma_device_type": 1 00:08:10.000 }, 00:08:10.000 { 00:08:10.000 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:10.000 "dma_device_type": 2 00:08:10.000 } 00:08:10.000 ], 00:08:10.000 "driver_specific": {} 00:08:10.000 } 00:08:10.000 ] 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.000 "name": "Existed_Raid", 00:08:10.000 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:10.000 "strip_size_kb": 64, 00:08:10.000 "state": "online", 00:08:10.000 "raid_level": "raid0", 00:08:10.000 "superblock": true, 00:08:10.000 "num_base_bdevs": 3, 00:08:10.000 "num_base_bdevs_discovered": 3, 00:08:10.000 "num_base_bdevs_operational": 3, 00:08:10.000 "base_bdevs_list": [ 00:08:10.000 { 00:08:10.000 "name": "NewBaseBdev", 00:08:10.000 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:10.000 "is_configured": true, 00:08:10.000 "data_offset": 2048, 00:08:10.000 "data_size": 63488 00:08:10.000 }, 00:08:10.000 { 00:08:10.000 "name": "BaseBdev2", 00:08:10.000 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:10.000 "is_configured": true, 00:08:10.000 "data_offset": 2048, 00:08:10.000 "data_size": 63488 00:08:10.000 }, 00:08:10.000 { 00:08:10.000 "name": "BaseBdev3", 00:08:10.000 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:10.000 "is_configured": true, 00:08:10.000 "data_offset": 2048, 00:08:10.000 "data_size": 63488 00:08:10.000 } 00:08:10.000 ] 00:08:10.000 }' 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.000 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:10.259 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:10.260 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:10.260 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.260 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.260 [2024-09-30 14:07:14.898203] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:10.519 14:07:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.519 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:10.519 "name": "Existed_Raid", 00:08:10.519 "aliases": [ 00:08:10.519 "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6" 00:08:10.519 ], 00:08:10.519 "product_name": "Raid Volume", 00:08:10.519 "block_size": 512, 00:08:10.519 "num_blocks": 190464, 00:08:10.519 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:10.519 "assigned_rate_limits": { 00:08:10.519 "rw_ios_per_sec": 0, 00:08:10.519 "rw_mbytes_per_sec": 0, 00:08:10.519 "r_mbytes_per_sec": 0, 00:08:10.519 "w_mbytes_per_sec": 0 00:08:10.519 }, 00:08:10.519 "claimed": false, 00:08:10.519 "zoned": false, 00:08:10.519 "supported_io_types": { 00:08:10.519 "read": true, 00:08:10.519 "write": true, 00:08:10.519 "unmap": true, 00:08:10.519 "flush": true, 00:08:10.519 "reset": true, 00:08:10.519 "nvme_admin": false, 00:08:10.519 "nvme_io": false, 00:08:10.519 "nvme_io_md": false, 00:08:10.519 "write_zeroes": true, 00:08:10.519 "zcopy": false, 00:08:10.519 "get_zone_info": false, 00:08:10.519 "zone_management": false, 00:08:10.519 "zone_append": false, 00:08:10.519 "compare": false, 00:08:10.519 "compare_and_write": false, 00:08:10.519 "abort": false, 00:08:10.519 "seek_hole": false, 00:08:10.519 "seek_data": false, 00:08:10.519 "copy": false, 00:08:10.519 "nvme_iov_md": false 00:08:10.519 }, 00:08:10.519 "memory_domains": [ 00:08:10.519 { 00:08:10.520 "dma_device_id": "system", 00:08:10.520 "dma_device_type": 1 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:10.520 "dma_device_type": 2 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "dma_device_id": "system", 00:08:10.520 "dma_device_type": 1 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:10.520 "dma_device_type": 2 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "dma_device_id": "system", 00:08:10.520 "dma_device_type": 1 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:10.520 "dma_device_type": 2 00:08:10.520 } 00:08:10.520 ], 00:08:10.520 "driver_specific": { 00:08:10.520 "raid": { 00:08:10.520 "uuid": "c3847346-d330-4ddc-8e65-d9dc4bd0b1a6", 00:08:10.520 "strip_size_kb": 64, 00:08:10.520 "state": "online", 00:08:10.520 "raid_level": "raid0", 00:08:10.520 "superblock": true, 00:08:10.520 "num_base_bdevs": 3, 00:08:10.520 "num_base_bdevs_discovered": 3, 00:08:10.520 "num_base_bdevs_operational": 3, 00:08:10.520 "base_bdevs_list": [ 00:08:10.520 { 00:08:10.520 "name": "NewBaseBdev", 00:08:10.520 "uuid": "6db25d93-61a3-4c6c-9049-c38bc8a63297", 00:08:10.520 "is_configured": true, 00:08:10.520 "data_offset": 2048, 00:08:10.520 "data_size": 63488 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "name": "BaseBdev2", 00:08:10.520 "uuid": "2730fcef-132e-4aa4-a51b-df351df08629", 00:08:10.520 "is_configured": true, 00:08:10.520 "data_offset": 2048, 00:08:10.520 "data_size": 63488 00:08:10.520 }, 00:08:10.520 { 00:08:10.520 "name": "BaseBdev3", 00:08:10.520 "uuid": "5c8647c6-87dd-484f-9a76-bb6d6fa3a985", 00:08:10.520 "is_configured": true, 00:08:10.520 "data_offset": 2048, 00:08:10.520 "data_size": 63488 00:08:10.520 } 00:08:10.520 ] 00:08:10.520 } 00:08:10.520 } 00:08:10.520 }' 00:08:10.520 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:10.520 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:10.520 BaseBdev2 00:08:10.520 BaseBdev3' 00:08:10.520 14:07:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:10.520 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.780 [2024-09-30 14:07:15.197401] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:10.780 [2024-09-30 14:07:15.197467] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:10.780 [2024-09-30 14:07:15.197543] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:10.780 [2024-09-30 14:07:15.197598] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:10.780 [2024-09-30 14:07:15.197607] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 76680 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 76680 ']' 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 76680 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76680 00:08:10.780 killing process with pid 76680 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76680' 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 76680 00:08:10.780 [2024-09-30 14:07:15.245235] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:10.780 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 76680 00:08:10.780 [2024-09-30 14:07:15.275603] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:11.040 14:07:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:11.040 ************************************ 00:08:11.040 END TEST raid_state_function_test_sb 00:08:11.040 ************************************ 00:08:11.040 00:08:11.040 real 0m8.678s 00:08:11.040 user 0m14.786s 00:08:11.040 sys 0m1.761s 00:08:11.040 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:11.041 14:07:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.041 14:07:15 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:11.041 14:07:15 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:11.041 14:07:15 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:11.041 14:07:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:11.041 ************************************ 00:08:11.041 START TEST raid_superblock_test 00:08:11.041 ************************************ 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 3 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77284 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77284 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 77284 ']' 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:11.041 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:11.041 14:07:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.041 [2024-09-30 14:07:15.678185] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:11.041 [2024-09-30 14:07:15.678387] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77284 ] 00:08:11.301 [2024-09-30 14:07:15.808747] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:11.301 [2024-09-30 14:07:15.838656] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:11.301 [2024-09-30 14:07:15.883679] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:11.301 [2024-09-30 14:07:15.924762] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.301 [2024-09-30 14:07:15.924804] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.870 malloc1 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.870 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.870 [2024-09-30 14:07:16.522131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:11.870 [2024-09-30 14:07:16.522254] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:11.870 [2024-09-30 14:07:16.522296] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:11.870 [2024-09-30 14:07:16.522346] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:11.870 [2024-09-30 14:07:16.524441] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:11.870 [2024-09-30 14:07:16.524524] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:12.130 pt1 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.130 malloc2 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.130 [2024-09-30 14:07:16.565977] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:12.130 [2024-09-30 14:07:16.566152] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.130 [2024-09-30 14:07:16.566196] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:12.130 [2024-09-30 14:07:16.566216] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.130 [2024-09-30 14:07:16.570675] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.130 [2024-09-30 14:07:16.570739] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:12.130 pt2 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:12.130 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.131 malloc3 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.131 [2024-09-30 14:07:16.596298] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:12.131 [2024-09-30 14:07:16.596383] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.131 [2024-09-30 14:07:16.596425] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:12.131 [2024-09-30 14:07:16.596453] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.131 [2024-09-30 14:07:16.598517] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.131 [2024-09-30 14:07:16.598588] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:12.131 pt3 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.131 [2024-09-30 14:07:16.608342] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:12.131 [2024-09-30 14:07:16.610180] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:12.131 [2024-09-30 14:07:16.610275] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:12.131 [2024-09-30 14:07:16.610444] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:12.131 [2024-09-30 14:07:16.610509] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:12.131 [2024-09-30 14:07:16.610776] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:12.131 [2024-09-30 14:07:16.610941] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:12.131 [2024-09-30 14:07:16.610986] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:08:12.131 [2024-09-30 14:07:16.611131] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.131 "name": "raid_bdev1", 00:08:12.131 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:12.131 "strip_size_kb": 64, 00:08:12.131 "state": "online", 00:08:12.131 "raid_level": "raid0", 00:08:12.131 "superblock": true, 00:08:12.131 "num_base_bdevs": 3, 00:08:12.131 "num_base_bdevs_discovered": 3, 00:08:12.131 "num_base_bdevs_operational": 3, 00:08:12.131 "base_bdevs_list": [ 00:08:12.131 { 00:08:12.131 "name": "pt1", 00:08:12.131 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.131 "is_configured": true, 00:08:12.131 "data_offset": 2048, 00:08:12.131 "data_size": 63488 00:08:12.131 }, 00:08:12.131 { 00:08:12.131 "name": "pt2", 00:08:12.131 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.131 "is_configured": true, 00:08:12.131 "data_offset": 2048, 00:08:12.131 "data_size": 63488 00:08:12.131 }, 00:08:12.131 { 00:08:12.131 "name": "pt3", 00:08:12.131 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.131 "is_configured": true, 00:08:12.131 "data_offset": 2048, 00:08:12.131 "data_size": 63488 00:08:12.131 } 00:08:12.131 ] 00:08:12.131 }' 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.131 14:07:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.391 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:12.391 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:12.391 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:12.391 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:12.391 14:07:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.391 [2024-09-30 14:07:17.011894] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:12.391 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.650 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:12.650 "name": "raid_bdev1", 00:08:12.650 "aliases": [ 00:08:12.650 "0a5676f3-8411-4e4d-b853-f27c7a8651e3" 00:08:12.650 ], 00:08:12.650 "product_name": "Raid Volume", 00:08:12.650 "block_size": 512, 00:08:12.650 "num_blocks": 190464, 00:08:12.650 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:12.650 "assigned_rate_limits": { 00:08:12.650 "rw_ios_per_sec": 0, 00:08:12.650 "rw_mbytes_per_sec": 0, 00:08:12.650 "r_mbytes_per_sec": 0, 00:08:12.650 "w_mbytes_per_sec": 0 00:08:12.650 }, 00:08:12.650 "claimed": false, 00:08:12.650 "zoned": false, 00:08:12.650 "supported_io_types": { 00:08:12.650 "read": true, 00:08:12.650 "write": true, 00:08:12.650 "unmap": true, 00:08:12.650 "flush": true, 00:08:12.650 "reset": true, 00:08:12.650 "nvme_admin": false, 00:08:12.650 "nvme_io": false, 00:08:12.650 "nvme_io_md": false, 00:08:12.650 "write_zeroes": true, 00:08:12.650 "zcopy": false, 00:08:12.650 "get_zone_info": false, 00:08:12.650 "zone_management": false, 00:08:12.650 "zone_append": false, 00:08:12.650 "compare": false, 00:08:12.650 "compare_and_write": false, 00:08:12.650 "abort": false, 00:08:12.650 "seek_hole": false, 00:08:12.650 "seek_data": false, 00:08:12.650 "copy": false, 00:08:12.650 "nvme_iov_md": false 00:08:12.650 }, 00:08:12.650 "memory_domains": [ 00:08:12.650 { 00:08:12.650 "dma_device_id": "system", 00:08:12.650 "dma_device_type": 1 00:08:12.650 }, 00:08:12.650 { 00:08:12.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.650 "dma_device_type": 2 00:08:12.650 }, 00:08:12.650 { 00:08:12.650 "dma_device_id": "system", 00:08:12.650 "dma_device_type": 1 00:08:12.650 }, 00:08:12.650 { 00:08:12.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.650 "dma_device_type": 2 00:08:12.650 }, 00:08:12.650 { 00:08:12.650 "dma_device_id": "system", 00:08:12.650 "dma_device_type": 1 00:08:12.650 }, 00:08:12.650 { 00:08:12.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.650 "dma_device_type": 2 00:08:12.650 } 00:08:12.650 ], 00:08:12.650 "driver_specific": { 00:08:12.650 "raid": { 00:08:12.650 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:12.650 "strip_size_kb": 64, 00:08:12.650 "state": "online", 00:08:12.650 "raid_level": "raid0", 00:08:12.650 "superblock": true, 00:08:12.651 "num_base_bdevs": 3, 00:08:12.651 "num_base_bdevs_discovered": 3, 00:08:12.651 "num_base_bdevs_operational": 3, 00:08:12.651 "base_bdevs_list": [ 00:08:12.651 { 00:08:12.651 "name": "pt1", 00:08:12.651 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.651 "is_configured": true, 00:08:12.651 "data_offset": 2048, 00:08:12.651 "data_size": 63488 00:08:12.651 }, 00:08:12.651 { 00:08:12.651 "name": "pt2", 00:08:12.651 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.651 "is_configured": true, 00:08:12.651 "data_offset": 2048, 00:08:12.651 "data_size": 63488 00:08:12.651 }, 00:08:12.651 { 00:08:12.651 "name": "pt3", 00:08:12.651 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.651 "is_configured": true, 00:08:12.651 "data_offset": 2048, 00:08:12.651 "data_size": 63488 00:08:12.651 } 00:08:12.651 ] 00:08:12.651 } 00:08:12.651 } 00:08:12.651 }' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:12.651 pt2 00:08:12.651 pt3' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.651 [2024-09-30 14:07:17.251417] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=0a5676f3-8411-4e4d-b853-f27c7a8651e3 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 0a5676f3-8411-4e4d-b853-f27c7a8651e3 ']' 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.651 [2024-09-30 14:07:17.299083] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:12.651 [2024-09-30 14:07:17.299107] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:12.651 [2024-09-30 14:07:17.299171] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:12.651 [2024-09-30 14:07:17.299227] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:12.651 [2024-09-30 14:07:17.299245] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:08:12.651 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:12.911 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.912 [2024-09-30 14:07:17.438881] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:12.912 [2024-09-30 14:07:17.440716] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:12.912 [2024-09-30 14:07:17.440764] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:12.912 [2024-09-30 14:07:17.440809] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:12.912 [2024-09-30 14:07:17.440849] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:12.912 [2024-09-30 14:07:17.440865] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:12.912 [2024-09-30 14:07:17.440878] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:12.912 [2024-09-30 14:07:17.440887] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:08:12.912 request: 00:08:12.912 { 00:08:12.912 "name": "raid_bdev1", 00:08:12.912 "raid_level": "raid0", 00:08:12.912 "base_bdevs": [ 00:08:12.912 "malloc1", 00:08:12.912 "malloc2", 00:08:12.912 "malloc3" 00:08:12.912 ], 00:08:12.912 "strip_size_kb": 64, 00:08:12.912 "superblock": false, 00:08:12.912 "method": "bdev_raid_create", 00:08:12.912 "req_id": 1 00:08:12.912 } 00:08:12.912 Got JSON-RPC error response 00:08:12.912 response: 00:08:12.912 { 00:08:12.912 "code": -17, 00:08:12.912 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:12.912 } 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.912 [2024-09-30 14:07:17.502731] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:12.912 [2024-09-30 14:07:17.502811] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.912 [2024-09-30 14:07:17.502845] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:12.912 [2024-09-30 14:07:17.502873] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.912 [2024-09-30 14:07:17.504916] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.912 [2024-09-30 14:07:17.504981] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:12.912 [2024-09-30 14:07:17.505062] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:12.912 [2024-09-30 14:07:17.505117] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:12.912 pt1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.912 "name": "raid_bdev1", 00:08:12.912 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:12.912 "strip_size_kb": 64, 00:08:12.912 "state": "configuring", 00:08:12.912 "raid_level": "raid0", 00:08:12.912 "superblock": true, 00:08:12.912 "num_base_bdevs": 3, 00:08:12.912 "num_base_bdevs_discovered": 1, 00:08:12.912 "num_base_bdevs_operational": 3, 00:08:12.912 "base_bdevs_list": [ 00:08:12.912 { 00:08:12.912 "name": "pt1", 00:08:12.912 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.912 "is_configured": true, 00:08:12.912 "data_offset": 2048, 00:08:12.912 "data_size": 63488 00:08:12.912 }, 00:08:12.912 { 00:08:12.912 "name": null, 00:08:12.912 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.912 "is_configured": false, 00:08:12.912 "data_offset": 2048, 00:08:12.912 "data_size": 63488 00:08:12.912 }, 00:08:12.912 { 00:08:12.912 "name": null, 00:08:12.912 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.912 "is_configured": false, 00:08:12.912 "data_offset": 2048, 00:08:12.912 "data_size": 63488 00:08:12.912 } 00:08:12.912 ] 00:08:12.912 }' 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.912 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.481 [2024-09-30 14:07:17.898070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:13.481 [2024-09-30 14:07:17.898177] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.481 [2024-09-30 14:07:17.898219] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:08:13.481 [2024-09-30 14:07:17.898247] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.481 [2024-09-30 14:07:17.898647] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.481 [2024-09-30 14:07:17.898702] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:13.481 [2024-09-30 14:07:17.898794] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:13.481 [2024-09-30 14:07:17.898838] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:13.481 pt2 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.481 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.482 [2024-09-30 14:07:17.910067] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.482 "name": "raid_bdev1", 00:08:13.482 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:13.482 "strip_size_kb": 64, 00:08:13.482 "state": "configuring", 00:08:13.482 "raid_level": "raid0", 00:08:13.482 "superblock": true, 00:08:13.482 "num_base_bdevs": 3, 00:08:13.482 "num_base_bdevs_discovered": 1, 00:08:13.482 "num_base_bdevs_operational": 3, 00:08:13.482 "base_bdevs_list": [ 00:08:13.482 { 00:08:13.482 "name": "pt1", 00:08:13.482 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:13.482 "is_configured": true, 00:08:13.482 "data_offset": 2048, 00:08:13.482 "data_size": 63488 00:08:13.482 }, 00:08:13.482 { 00:08:13.482 "name": null, 00:08:13.482 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:13.482 "is_configured": false, 00:08:13.482 "data_offset": 0, 00:08:13.482 "data_size": 63488 00:08:13.482 }, 00:08:13.482 { 00:08:13.482 "name": null, 00:08:13.482 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:13.482 "is_configured": false, 00:08:13.482 "data_offset": 2048, 00:08:13.482 "data_size": 63488 00:08:13.482 } 00:08:13.482 ] 00:08:13.482 }' 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.482 14:07:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.742 [2024-09-30 14:07:18.321303] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:13.742 [2024-09-30 14:07:18.321394] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.742 [2024-09-30 14:07:18.321424] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:13.742 [2024-09-30 14:07:18.321453] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.742 [2024-09-30 14:07:18.321824] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.742 [2024-09-30 14:07:18.321880] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:13.742 [2024-09-30 14:07:18.321970] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:13.742 [2024-09-30 14:07:18.322017] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:13.742 pt2 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.742 [2024-09-30 14:07:18.333290] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:13.742 [2024-09-30 14:07:18.333372] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.742 [2024-09-30 14:07:18.333401] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:13.742 [2024-09-30 14:07:18.333438] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.742 [2024-09-30 14:07:18.333764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.742 [2024-09-30 14:07:18.333819] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:13.742 [2024-09-30 14:07:18.333895] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:13.742 [2024-09-30 14:07:18.333942] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:13.742 [2024-09-30 14:07:18.334039] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:13.742 [2024-09-30 14:07:18.334076] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:13.742 [2024-09-30 14:07:18.334300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:08:13.742 [2024-09-30 14:07:18.334439] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:13.742 [2024-09-30 14:07:18.334496] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:08:13.742 [2024-09-30 14:07:18.334653] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:13.742 pt3 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.742 "name": "raid_bdev1", 00:08:13.742 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:13.742 "strip_size_kb": 64, 00:08:13.742 "state": "online", 00:08:13.742 "raid_level": "raid0", 00:08:13.742 "superblock": true, 00:08:13.742 "num_base_bdevs": 3, 00:08:13.742 "num_base_bdevs_discovered": 3, 00:08:13.742 "num_base_bdevs_operational": 3, 00:08:13.742 "base_bdevs_list": [ 00:08:13.742 { 00:08:13.742 "name": "pt1", 00:08:13.742 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:13.742 "is_configured": true, 00:08:13.742 "data_offset": 2048, 00:08:13.742 "data_size": 63488 00:08:13.742 }, 00:08:13.742 { 00:08:13.742 "name": "pt2", 00:08:13.742 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:13.742 "is_configured": true, 00:08:13.742 "data_offset": 2048, 00:08:13.742 "data_size": 63488 00:08:13.742 }, 00:08:13.742 { 00:08:13.742 "name": "pt3", 00:08:13.742 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:13.742 "is_configured": true, 00:08:13.742 "data_offset": 2048, 00:08:13.742 "data_size": 63488 00:08:13.742 } 00:08:13.742 ] 00:08:13.742 }' 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.742 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.383 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 [2024-09-30 14:07:18.784854] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:14.384 "name": "raid_bdev1", 00:08:14.384 "aliases": [ 00:08:14.384 "0a5676f3-8411-4e4d-b853-f27c7a8651e3" 00:08:14.384 ], 00:08:14.384 "product_name": "Raid Volume", 00:08:14.384 "block_size": 512, 00:08:14.384 "num_blocks": 190464, 00:08:14.384 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:14.384 "assigned_rate_limits": { 00:08:14.384 "rw_ios_per_sec": 0, 00:08:14.384 "rw_mbytes_per_sec": 0, 00:08:14.384 "r_mbytes_per_sec": 0, 00:08:14.384 "w_mbytes_per_sec": 0 00:08:14.384 }, 00:08:14.384 "claimed": false, 00:08:14.384 "zoned": false, 00:08:14.384 "supported_io_types": { 00:08:14.384 "read": true, 00:08:14.384 "write": true, 00:08:14.384 "unmap": true, 00:08:14.384 "flush": true, 00:08:14.384 "reset": true, 00:08:14.384 "nvme_admin": false, 00:08:14.384 "nvme_io": false, 00:08:14.384 "nvme_io_md": false, 00:08:14.384 "write_zeroes": true, 00:08:14.384 "zcopy": false, 00:08:14.384 "get_zone_info": false, 00:08:14.384 "zone_management": false, 00:08:14.384 "zone_append": false, 00:08:14.384 "compare": false, 00:08:14.384 "compare_and_write": false, 00:08:14.384 "abort": false, 00:08:14.384 "seek_hole": false, 00:08:14.384 "seek_data": false, 00:08:14.384 "copy": false, 00:08:14.384 "nvme_iov_md": false 00:08:14.384 }, 00:08:14.384 "memory_domains": [ 00:08:14.384 { 00:08:14.384 "dma_device_id": "system", 00:08:14.384 "dma_device_type": 1 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.384 "dma_device_type": 2 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "dma_device_id": "system", 00:08:14.384 "dma_device_type": 1 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.384 "dma_device_type": 2 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "dma_device_id": "system", 00:08:14.384 "dma_device_type": 1 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.384 "dma_device_type": 2 00:08:14.384 } 00:08:14.384 ], 00:08:14.384 "driver_specific": { 00:08:14.384 "raid": { 00:08:14.384 "uuid": "0a5676f3-8411-4e4d-b853-f27c7a8651e3", 00:08:14.384 "strip_size_kb": 64, 00:08:14.384 "state": "online", 00:08:14.384 "raid_level": "raid0", 00:08:14.384 "superblock": true, 00:08:14.384 "num_base_bdevs": 3, 00:08:14.384 "num_base_bdevs_discovered": 3, 00:08:14.384 "num_base_bdevs_operational": 3, 00:08:14.384 "base_bdevs_list": [ 00:08:14.384 { 00:08:14.384 "name": "pt1", 00:08:14.384 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:14.384 "is_configured": true, 00:08:14.384 "data_offset": 2048, 00:08:14.384 "data_size": 63488 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "name": "pt2", 00:08:14.384 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:14.384 "is_configured": true, 00:08:14.384 "data_offset": 2048, 00:08:14.384 "data_size": 63488 00:08:14.384 }, 00:08:14.384 { 00:08:14.384 "name": "pt3", 00:08:14.384 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:14.384 "is_configured": true, 00:08:14.384 "data_offset": 2048, 00:08:14.384 "data_size": 63488 00:08:14.384 } 00:08:14.384 ] 00:08:14.384 } 00:08:14.384 } 00:08:14.384 }' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:14.384 pt2 00:08:14.384 pt3' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.384 14:07:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.384 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 [2024-09-30 14:07:19.020356] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 0a5676f3-8411-4e4d-b853-f27c7a8651e3 '!=' 0a5676f3-8411-4e4d-b853-f27c7a8651e3 ']' 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77284 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 77284 ']' 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 77284 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77284 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:14.644 killing process with pid 77284 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77284' 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 77284 00:08:14.644 [2024-09-30 14:07:19.104905] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:14.644 [2024-09-30 14:07:19.104991] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:14.644 [2024-09-30 14:07:19.105048] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:14.644 [2024-09-30 14:07:19.105059] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:08:14.644 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 77284 00:08:14.644 [2024-09-30 14:07:19.137668] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:14.904 14:07:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:14.904 00:08:14.904 real 0m3.784s 00:08:14.904 user 0m5.873s 00:08:14.904 sys 0m0.869s 00:08:14.904 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:14.904 14:07:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.904 ************************************ 00:08:14.904 END TEST raid_superblock_test 00:08:14.904 ************************************ 00:08:14.904 14:07:19 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:14.904 14:07:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:14.904 14:07:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:14.904 14:07:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:14.904 ************************************ 00:08:14.904 START TEST raid_read_error_test 00:08:14.904 ************************************ 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 read 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.8FzGsE7B6G 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77520 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77520 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 77520 ']' 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:14.904 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:14.904 14:07:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.904 [2024-09-30 14:07:19.549802] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:14.904 [2024-09-30 14:07:19.550005] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77520 ] 00:08:15.164 [2024-09-30 14:07:19.680550] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:15.164 [2024-09-30 14:07:19.710837] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:15.164 [2024-09-30 14:07:19.755969] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:15.164 [2024-09-30 14:07:19.797168] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.164 [2024-09-30 14:07:19.797302] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.734 BaseBdev1_malloc 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.734 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 true 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 [2024-09-30 14:07:20.402933] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:15.994 [2024-09-30 14:07:20.402986] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.994 [2024-09-30 14:07:20.403003] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:15.994 [2024-09-30 14:07:20.403022] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.994 [2024-09-30 14:07:20.405106] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.994 [2024-09-30 14:07:20.405179] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:15.994 BaseBdev1 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 BaseBdev2_malloc 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 true 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 [2024-09-30 14:07:20.456131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:15.994 [2024-09-30 14:07:20.456223] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.994 [2024-09-30 14:07:20.456243] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:15.994 [2024-09-30 14:07:20.456254] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.994 [2024-09-30 14:07:20.458292] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.994 [2024-09-30 14:07:20.458335] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:15.994 BaseBdev2 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.994 BaseBdev3_malloc 00:08:15.994 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.995 true 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.995 [2024-09-30 14:07:20.496413] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:15.995 [2024-09-30 14:07:20.496458] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.995 [2024-09-30 14:07:20.496473] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:15.995 [2024-09-30 14:07:20.496501] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.995 [2024-09-30 14:07:20.498510] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.995 [2024-09-30 14:07:20.498543] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:15.995 BaseBdev3 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.995 [2024-09-30 14:07:20.508461] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:15.995 [2024-09-30 14:07:20.510171] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:15.995 [2024-09-30 14:07:20.510245] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:15.995 [2024-09-30 14:07:20.510401] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:15.995 [2024-09-30 14:07:20.510411] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:15.995 [2024-09-30 14:07:20.510688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:15.995 [2024-09-30 14:07:20.510822] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:15.995 [2024-09-30 14:07:20.510843] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:08:15.995 [2024-09-30 14:07:20.510952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.995 "name": "raid_bdev1", 00:08:15.995 "uuid": "9bb9044f-1d74-465e-a1c5-1bd0c5c329be", 00:08:15.995 "strip_size_kb": 64, 00:08:15.995 "state": "online", 00:08:15.995 "raid_level": "raid0", 00:08:15.995 "superblock": true, 00:08:15.995 "num_base_bdevs": 3, 00:08:15.995 "num_base_bdevs_discovered": 3, 00:08:15.995 "num_base_bdevs_operational": 3, 00:08:15.995 "base_bdevs_list": [ 00:08:15.995 { 00:08:15.995 "name": "BaseBdev1", 00:08:15.995 "uuid": "33416e54-8d74-5ccb-a4fc-1a8d76db6ee0", 00:08:15.995 "is_configured": true, 00:08:15.995 "data_offset": 2048, 00:08:15.995 "data_size": 63488 00:08:15.995 }, 00:08:15.995 { 00:08:15.995 "name": "BaseBdev2", 00:08:15.995 "uuid": "09d078da-47f3-58a6-81e7-34bba3d98c20", 00:08:15.995 "is_configured": true, 00:08:15.995 "data_offset": 2048, 00:08:15.995 "data_size": 63488 00:08:15.995 }, 00:08:15.995 { 00:08:15.995 "name": "BaseBdev3", 00:08:15.995 "uuid": "a912da87-1e08-56ac-b91e-df6319bcf4c6", 00:08:15.995 "is_configured": true, 00:08:15.995 "data_offset": 2048, 00:08:15.995 "data_size": 63488 00:08:15.995 } 00:08:15.995 ] 00:08:15.995 }' 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.995 14:07:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.562 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:16.562 14:07:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:16.562 [2024-09-30 14:07:21.039941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.500 14:07:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.500 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.500 "name": "raid_bdev1", 00:08:17.500 "uuid": "9bb9044f-1d74-465e-a1c5-1bd0c5c329be", 00:08:17.500 "strip_size_kb": 64, 00:08:17.500 "state": "online", 00:08:17.500 "raid_level": "raid0", 00:08:17.500 "superblock": true, 00:08:17.500 "num_base_bdevs": 3, 00:08:17.500 "num_base_bdevs_discovered": 3, 00:08:17.500 "num_base_bdevs_operational": 3, 00:08:17.500 "base_bdevs_list": [ 00:08:17.500 { 00:08:17.500 "name": "BaseBdev1", 00:08:17.500 "uuid": "33416e54-8d74-5ccb-a4fc-1a8d76db6ee0", 00:08:17.500 "is_configured": true, 00:08:17.500 "data_offset": 2048, 00:08:17.500 "data_size": 63488 00:08:17.500 }, 00:08:17.500 { 00:08:17.500 "name": "BaseBdev2", 00:08:17.500 "uuid": "09d078da-47f3-58a6-81e7-34bba3d98c20", 00:08:17.500 "is_configured": true, 00:08:17.500 "data_offset": 2048, 00:08:17.500 "data_size": 63488 00:08:17.500 }, 00:08:17.500 { 00:08:17.500 "name": "BaseBdev3", 00:08:17.500 "uuid": "a912da87-1e08-56ac-b91e-df6319bcf4c6", 00:08:17.500 "is_configured": true, 00:08:17.500 "data_offset": 2048, 00:08:17.500 "data_size": 63488 00:08:17.500 } 00:08:17.500 ] 00:08:17.500 }' 00:08:17.500 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.500 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.758 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:17.758 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.758 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.758 [2024-09-30 14:07:22.395285] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:17.758 [2024-09-30 14:07:22.395392] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:17.758 [2024-09-30 14:07:22.397836] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.758 [2024-09-30 14:07:22.397875] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:17.758 [2024-09-30 14:07:22.397909] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.759 [2024-09-30 14:07:22.397918] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:08:17.759 { 00:08:17.759 "results": [ 00:08:17.759 { 00:08:17.759 "job": "raid_bdev1", 00:08:17.759 "core_mask": "0x1", 00:08:17.759 "workload": "randrw", 00:08:17.759 "percentage": 50, 00:08:17.759 "status": "finished", 00:08:17.759 "queue_depth": 1, 00:08:17.759 "io_size": 131072, 00:08:17.759 "runtime": 1.356362, 00:08:17.759 "iops": 17420.12825484642, 00:08:17.759 "mibps": 2177.5160318558023, 00:08:17.759 "io_failed": 1, 00:08:17.759 "io_timeout": 0, 00:08:17.759 "avg_latency_us": 79.57833592464002, 00:08:17.759 "min_latency_us": 18.668995633187773, 00:08:17.759 "max_latency_us": 1316.4436681222708 00:08:17.759 } 00:08:17.759 ], 00:08:17.759 "core_count": 1 00:08:17.759 } 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77520 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 77520 ']' 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 77520 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:17.759 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77520 00:08:18.017 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:18.017 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:18.017 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77520' 00:08:18.017 killing process with pid 77520 00:08:18.017 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 77520 00:08:18.017 [2024-09-30 14:07:22.433560] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:18.017 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 77520 00:08:18.017 [2024-09-30 14:07:22.458244] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.8FzGsE7B6G 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:18.277 00:08:18.277 real 0m3.255s 00:08:18.277 user 0m4.074s 00:08:18.277 sys 0m0.543s 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:18.277 14:07:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.277 ************************************ 00:08:18.277 END TEST raid_read_error_test 00:08:18.277 ************************************ 00:08:18.277 14:07:22 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:18.277 14:07:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:18.277 14:07:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:18.277 14:07:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:18.277 ************************************ 00:08:18.277 START TEST raid_write_error_test 00:08:18.277 ************************************ 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 write 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.zXqWEYaSvS 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77654 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77654 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 77654 ']' 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:18.277 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:18.277 14:07:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.277 [2024-09-30 14:07:22.876994] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:18.277 [2024-09-30 14:07:22.877190] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77654 ] 00:08:18.537 [2024-09-30 14:07:23.007642] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:18.537 [2024-09-30 14:07:23.037083] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.537 [2024-09-30 14:07:23.082403] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.537 [2024-09-30 14:07:23.124436] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.537 [2024-09-30 14:07:23.124604] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.105 BaseBdev1_malloc 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.105 true 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.105 [2024-09-30 14:07:23.730510] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:19.105 [2024-09-30 14:07:23.730605] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.105 [2024-09-30 14:07:23.730662] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:19.105 [2024-09-30 14:07:23.730695] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.105 [2024-09-30 14:07:23.732717] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.105 [2024-09-30 14:07:23.732783] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:19.105 BaseBdev1 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.105 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 BaseBdev2_malloc 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 true 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 [2024-09-30 14:07:23.788401] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:19.364 [2024-09-30 14:07:23.788548] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.364 [2024-09-30 14:07:23.788605] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:19.364 [2024-09-30 14:07:23.788708] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.364 [2024-09-30 14:07:23.791972] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.364 [2024-09-30 14:07:23.792080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:19.364 BaseBdev2 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 BaseBdev3_malloc 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 true 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 [2024-09-30 14:07:23.829049] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:19.364 [2024-09-30 14:07:23.829095] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.364 [2024-09-30 14:07:23.829110] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:19.364 [2024-09-30 14:07:23.829120] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.364 [2024-09-30 14:07:23.831212] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.364 [2024-09-30 14:07:23.831285] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:19.364 BaseBdev3 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.364 [2024-09-30 14:07:23.841094] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:19.364 [2024-09-30 14:07:23.842910] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:19.364 [2024-09-30 14:07:23.842985] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:19.364 [2024-09-30 14:07:23.843156] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:19.364 [2024-09-30 14:07:23.843167] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:19.364 [2024-09-30 14:07:23.843391] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:19.364 [2024-09-30 14:07:23.843531] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:19.364 [2024-09-30 14:07:23.843545] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:08:19.364 [2024-09-30 14:07:23.843686] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.364 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.365 "name": "raid_bdev1", 00:08:19.365 "uuid": "05c83e6f-8796-4dcc-b985-bd585d1dd04d", 00:08:19.365 "strip_size_kb": 64, 00:08:19.365 "state": "online", 00:08:19.365 "raid_level": "raid0", 00:08:19.365 "superblock": true, 00:08:19.365 "num_base_bdevs": 3, 00:08:19.365 "num_base_bdevs_discovered": 3, 00:08:19.365 "num_base_bdevs_operational": 3, 00:08:19.365 "base_bdevs_list": [ 00:08:19.365 { 00:08:19.365 "name": "BaseBdev1", 00:08:19.365 "uuid": "3880fbc0-cb7a-567c-b7ee-830af02f5ebd", 00:08:19.365 "is_configured": true, 00:08:19.365 "data_offset": 2048, 00:08:19.365 "data_size": 63488 00:08:19.365 }, 00:08:19.365 { 00:08:19.365 "name": "BaseBdev2", 00:08:19.365 "uuid": "0f541df9-c6f3-58c7-a57a-45d04c31d8b0", 00:08:19.365 "is_configured": true, 00:08:19.365 "data_offset": 2048, 00:08:19.365 "data_size": 63488 00:08:19.365 }, 00:08:19.365 { 00:08:19.365 "name": "BaseBdev3", 00:08:19.365 "uuid": "765e8b5b-8bfa-51a9-a121-4dfee021c83a", 00:08:19.365 "is_configured": true, 00:08:19.365 "data_offset": 2048, 00:08:19.365 "data_size": 63488 00:08:19.365 } 00:08:19.365 ] 00:08:19.365 }' 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.365 14:07:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.932 14:07:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:19.932 14:07:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:19.932 [2024-09-30 14:07:24.388474] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:20.870 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.871 "name": "raid_bdev1", 00:08:20.871 "uuid": "05c83e6f-8796-4dcc-b985-bd585d1dd04d", 00:08:20.871 "strip_size_kb": 64, 00:08:20.871 "state": "online", 00:08:20.871 "raid_level": "raid0", 00:08:20.871 "superblock": true, 00:08:20.871 "num_base_bdevs": 3, 00:08:20.871 "num_base_bdevs_discovered": 3, 00:08:20.871 "num_base_bdevs_operational": 3, 00:08:20.871 "base_bdevs_list": [ 00:08:20.871 { 00:08:20.871 "name": "BaseBdev1", 00:08:20.871 "uuid": "3880fbc0-cb7a-567c-b7ee-830af02f5ebd", 00:08:20.871 "is_configured": true, 00:08:20.871 "data_offset": 2048, 00:08:20.871 "data_size": 63488 00:08:20.871 }, 00:08:20.871 { 00:08:20.871 "name": "BaseBdev2", 00:08:20.871 "uuid": "0f541df9-c6f3-58c7-a57a-45d04c31d8b0", 00:08:20.871 "is_configured": true, 00:08:20.871 "data_offset": 2048, 00:08:20.871 "data_size": 63488 00:08:20.871 }, 00:08:20.871 { 00:08:20.871 "name": "BaseBdev3", 00:08:20.871 "uuid": "765e8b5b-8bfa-51a9-a121-4dfee021c83a", 00:08:20.871 "is_configured": true, 00:08:20.871 "data_offset": 2048, 00:08:20.871 "data_size": 63488 00:08:20.871 } 00:08:20.871 ] 00:08:20.871 }' 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.871 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.130 [2024-09-30 14:07:25.735463] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:21.130 [2024-09-30 14:07:25.735526] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:21.130 [2024-09-30 14:07:25.737928] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:21.130 [2024-09-30 14:07:25.737968] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:21.130 [2024-09-30 14:07:25.738002] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:21.130 [2024-09-30 14:07:25.738011] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:08:21.130 { 00:08:21.130 "results": [ 00:08:21.130 { 00:08:21.130 "job": "raid_bdev1", 00:08:21.130 "core_mask": "0x1", 00:08:21.130 "workload": "randrw", 00:08:21.130 "percentage": 50, 00:08:21.130 "status": "finished", 00:08:21.130 "queue_depth": 1, 00:08:21.130 "io_size": 131072, 00:08:21.130 "runtime": 1.347966, 00:08:21.130 "iops": 17438.867152435596, 00:08:21.130 "mibps": 2179.8583940544495, 00:08:21.130 "io_failed": 1, 00:08:21.130 "io_timeout": 0, 00:08:21.130 "avg_latency_us": 79.45029732515104, 00:08:21.130 "min_latency_us": 24.593886462882097, 00:08:21.130 "max_latency_us": 1345.0620087336245 00:08:21.130 } 00:08:21.130 ], 00:08:21.130 "core_count": 1 00:08:21.130 } 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77654 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 77654 ']' 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 77654 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77654 00:08:21.130 killing process with pid 77654 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77654' 00:08:21.130 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 77654 00:08:21.130 [2024-09-30 14:07:25.770128] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:21.131 14:07:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 77654 00:08:21.390 [2024-09-30 14:07:25.795146] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.zXqWEYaSvS 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:21.390 00:08:21.390 real 0m3.265s 00:08:21.390 user 0m4.106s 00:08:21.390 sys 0m0.528s 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:21.390 ************************************ 00:08:21.390 END TEST raid_write_error_test 00:08:21.390 ************************************ 00:08:21.390 14:07:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.651 14:07:26 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:21.651 14:07:26 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:21.651 14:07:26 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:21.651 14:07:26 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:21.651 14:07:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:21.651 ************************************ 00:08:21.651 START TEST raid_state_function_test 00:08:21.651 ************************************ 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 false 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:21.651 Process raid pid: 77782 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=77782 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77782' 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 77782 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 77782 ']' 00:08:21.651 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:21.651 14:07:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.651 [2024-09-30 14:07:26.204005] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:21.651 [2024-09-30 14:07:26.204107] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:21.911 [2024-09-30 14:07:26.336567] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:21.911 [2024-09-30 14:07:26.367767] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:21.911 [2024-09-30 14:07:26.413133] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:21.911 [2024-09-30 14:07:26.454089] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:21.911 [2024-09-30 14:07:26.454135] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:22.479 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.480 [2024-09-30 14:07:27.046833] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:22.480 [2024-09-30 14:07:27.046887] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:22.480 [2024-09-30 14:07:27.046903] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:22.480 [2024-09-30 14:07:27.046911] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:22.480 [2024-09-30 14:07:27.046921] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:22.480 [2024-09-30 14:07:27.046927] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.480 "name": "Existed_Raid", 00:08:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.480 "strip_size_kb": 64, 00:08:22.480 "state": "configuring", 00:08:22.480 "raid_level": "concat", 00:08:22.480 "superblock": false, 00:08:22.480 "num_base_bdevs": 3, 00:08:22.480 "num_base_bdevs_discovered": 0, 00:08:22.480 "num_base_bdevs_operational": 3, 00:08:22.480 "base_bdevs_list": [ 00:08:22.480 { 00:08:22.480 "name": "BaseBdev1", 00:08:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.480 "is_configured": false, 00:08:22.480 "data_offset": 0, 00:08:22.480 "data_size": 0 00:08:22.480 }, 00:08:22.480 { 00:08:22.480 "name": "BaseBdev2", 00:08:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.480 "is_configured": false, 00:08:22.480 "data_offset": 0, 00:08:22.480 "data_size": 0 00:08:22.480 }, 00:08:22.480 { 00:08:22.480 "name": "BaseBdev3", 00:08:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.480 "is_configured": false, 00:08:22.480 "data_offset": 0, 00:08:22.480 "data_size": 0 00:08:22.480 } 00:08:22.480 ] 00:08:22.480 }' 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.480 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 [2024-09-30 14:07:27.478048] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:23.049 [2024-09-30 14:07:27.478080] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 [2024-09-30 14:07:27.490030] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:23.049 [2024-09-30 14:07:27.490066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:23.049 [2024-09-30 14:07:27.490076] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:23.049 [2024-09-30 14:07:27.490083] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:23.049 [2024-09-30 14:07:27.490090] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:23.049 [2024-09-30 14:07:27.490096] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 [2024-09-30 14:07:27.510540] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:23.049 BaseBdev1 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.049 [ 00:08:23.049 { 00:08:23.049 "name": "BaseBdev1", 00:08:23.049 "aliases": [ 00:08:23.049 "2bfd28d1-07e6-42ef-975d-095b0c81ad29" 00:08:23.049 ], 00:08:23.049 "product_name": "Malloc disk", 00:08:23.049 "block_size": 512, 00:08:23.049 "num_blocks": 65536, 00:08:23.049 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:23.049 "assigned_rate_limits": { 00:08:23.049 "rw_ios_per_sec": 0, 00:08:23.049 "rw_mbytes_per_sec": 0, 00:08:23.049 "r_mbytes_per_sec": 0, 00:08:23.049 "w_mbytes_per_sec": 0 00:08:23.049 }, 00:08:23.049 "claimed": true, 00:08:23.049 "claim_type": "exclusive_write", 00:08:23.049 "zoned": false, 00:08:23.049 "supported_io_types": { 00:08:23.049 "read": true, 00:08:23.049 "write": true, 00:08:23.049 "unmap": true, 00:08:23.049 "flush": true, 00:08:23.049 "reset": true, 00:08:23.049 "nvme_admin": false, 00:08:23.049 "nvme_io": false, 00:08:23.049 "nvme_io_md": false, 00:08:23.049 "write_zeroes": true, 00:08:23.049 "zcopy": true, 00:08:23.049 "get_zone_info": false, 00:08:23.049 "zone_management": false, 00:08:23.049 "zone_append": false, 00:08:23.049 "compare": false, 00:08:23.049 "compare_and_write": false, 00:08:23.049 "abort": true, 00:08:23.049 "seek_hole": false, 00:08:23.049 "seek_data": false, 00:08:23.049 "copy": true, 00:08:23.049 "nvme_iov_md": false 00:08:23.049 }, 00:08:23.049 "memory_domains": [ 00:08:23.049 { 00:08:23.049 "dma_device_id": "system", 00:08:23.049 "dma_device_type": 1 00:08:23.049 }, 00:08:23.049 { 00:08:23.049 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.049 "dma_device_type": 2 00:08:23.049 } 00:08:23.049 ], 00:08:23.049 "driver_specific": {} 00:08:23.049 } 00:08:23.049 ] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.049 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.050 "name": "Existed_Raid", 00:08:23.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.050 "strip_size_kb": 64, 00:08:23.050 "state": "configuring", 00:08:23.050 "raid_level": "concat", 00:08:23.050 "superblock": false, 00:08:23.050 "num_base_bdevs": 3, 00:08:23.050 "num_base_bdevs_discovered": 1, 00:08:23.050 "num_base_bdevs_operational": 3, 00:08:23.050 "base_bdevs_list": [ 00:08:23.050 { 00:08:23.050 "name": "BaseBdev1", 00:08:23.050 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:23.050 "is_configured": true, 00:08:23.050 "data_offset": 0, 00:08:23.050 "data_size": 65536 00:08:23.050 }, 00:08:23.050 { 00:08:23.050 "name": "BaseBdev2", 00:08:23.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.050 "is_configured": false, 00:08:23.050 "data_offset": 0, 00:08:23.050 "data_size": 0 00:08:23.050 }, 00:08:23.050 { 00:08:23.050 "name": "BaseBdev3", 00:08:23.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.050 "is_configured": false, 00:08:23.050 "data_offset": 0, 00:08:23.050 "data_size": 0 00:08:23.050 } 00:08:23.050 ] 00:08:23.050 }' 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.050 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.309 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:23.309 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.309 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.568 [2024-09-30 14:07:27.969777] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:23.568 [2024-09-30 14:07:27.969886] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.568 [2024-09-30 14:07:27.981764] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:23.568 [2024-09-30 14:07:27.983624] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:23.568 [2024-09-30 14:07:27.983695] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:23.568 [2024-09-30 14:07:27.983726] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:23.568 [2024-09-30 14:07:27.983747] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.568 14:07:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.568 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.568 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.568 "name": "Existed_Raid", 00:08:23.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.568 "strip_size_kb": 64, 00:08:23.568 "state": "configuring", 00:08:23.568 "raid_level": "concat", 00:08:23.568 "superblock": false, 00:08:23.568 "num_base_bdevs": 3, 00:08:23.568 "num_base_bdevs_discovered": 1, 00:08:23.568 "num_base_bdevs_operational": 3, 00:08:23.568 "base_bdevs_list": [ 00:08:23.568 { 00:08:23.568 "name": "BaseBdev1", 00:08:23.568 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:23.568 "is_configured": true, 00:08:23.568 "data_offset": 0, 00:08:23.568 "data_size": 65536 00:08:23.568 }, 00:08:23.568 { 00:08:23.568 "name": "BaseBdev2", 00:08:23.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.568 "is_configured": false, 00:08:23.568 "data_offset": 0, 00:08:23.568 "data_size": 0 00:08:23.568 }, 00:08:23.568 { 00:08:23.568 "name": "BaseBdev3", 00:08:23.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.568 "is_configured": false, 00:08:23.568 "data_offset": 0, 00:08:23.568 "data_size": 0 00:08:23.568 } 00:08:23.568 ] 00:08:23.568 }' 00:08:23.568 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.568 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.832 [2024-09-30 14:07:28.469079] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:23.832 BaseBdev2 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.832 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.091 [ 00:08:24.091 { 00:08:24.091 "name": "BaseBdev2", 00:08:24.091 "aliases": [ 00:08:24.091 "843df0d6-a8b8-498e-baea-3e2a90828a89" 00:08:24.091 ], 00:08:24.091 "product_name": "Malloc disk", 00:08:24.091 "block_size": 512, 00:08:24.091 "num_blocks": 65536, 00:08:24.091 "uuid": "843df0d6-a8b8-498e-baea-3e2a90828a89", 00:08:24.091 "assigned_rate_limits": { 00:08:24.091 "rw_ios_per_sec": 0, 00:08:24.091 "rw_mbytes_per_sec": 0, 00:08:24.091 "r_mbytes_per_sec": 0, 00:08:24.091 "w_mbytes_per_sec": 0 00:08:24.091 }, 00:08:24.091 "claimed": true, 00:08:24.091 "claim_type": "exclusive_write", 00:08:24.091 "zoned": false, 00:08:24.091 "supported_io_types": { 00:08:24.091 "read": true, 00:08:24.091 "write": true, 00:08:24.091 "unmap": true, 00:08:24.091 "flush": true, 00:08:24.091 "reset": true, 00:08:24.091 "nvme_admin": false, 00:08:24.091 "nvme_io": false, 00:08:24.091 "nvme_io_md": false, 00:08:24.091 "write_zeroes": true, 00:08:24.091 "zcopy": true, 00:08:24.091 "get_zone_info": false, 00:08:24.091 "zone_management": false, 00:08:24.091 "zone_append": false, 00:08:24.091 "compare": false, 00:08:24.091 "compare_and_write": false, 00:08:24.091 "abort": true, 00:08:24.091 "seek_hole": false, 00:08:24.091 "seek_data": false, 00:08:24.091 "copy": true, 00:08:24.091 "nvme_iov_md": false 00:08:24.091 }, 00:08:24.091 "memory_domains": [ 00:08:24.091 { 00:08:24.091 "dma_device_id": "system", 00:08:24.091 "dma_device_type": 1 00:08:24.091 }, 00:08:24.091 { 00:08:24.091 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.091 "dma_device_type": 2 00:08:24.091 } 00:08:24.091 ], 00:08:24.091 "driver_specific": {} 00:08:24.091 } 00:08:24.091 ] 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.091 "name": "Existed_Raid", 00:08:24.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.091 "strip_size_kb": 64, 00:08:24.091 "state": "configuring", 00:08:24.091 "raid_level": "concat", 00:08:24.091 "superblock": false, 00:08:24.091 "num_base_bdevs": 3, 00:08:24.091 "num_base_bdevs_discovered": 2, 00:08:24.091 "num_base_bdevs_operational": 3, 00:08:24.091 "base_bdevs_list": [ 00:08:24.091 { 00:08:24.091 "name": "BaseBdev1", 00:08:24.091 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:24.091 "is_configured": true, 00:08:24.091 "data_offset": 0, 00:08:24.091 "data_size": 65536 00:08:24.091 }, 00:08:24.091 { 00:08:24.091 "name": "BaseBdev2", 00:08:24.091 "uuid": "843df0d6-a8b8-498e-baea-3e2a90828a89", 00:08:24.091 "is_configured": true, 00:08:24.091 "data_offset": 0, 00:08:24.091 "data_size": 65536 00:08:24.091 }, 00:08:24.091 { 00:08:24.091 "name": "BaseBdev3", 00:08:24.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.091 "is_configured": false, 00:08:24.091 "data_offset": 0, 00:08:24.091 "data_size": 0 00:08:24.091 } 00:08:24.091 ] 00:08:24.091 }' 00:08:24.091 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.092 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.350 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:24.350 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.350 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.350 [2024-09-30 14:07:28.987067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:24.350 [2024-09-30 14:07:28.987182] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:24.350 [2024-09-30 14:07:28.987194] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:24.350 [2024-09-30 14:07:28.987496] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:24.351 [2024-09-30 14:07:28.987630] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:24.351 [2024-09-30 14:07:28.987646] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:24.351 [2024-09-30 14:07:28.987846] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:24.351 BaseBdev3 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.351 14:07:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.608 [ 00:08:24.608 { 00:08:24.608 "name": "BaseBdev3", 00:08:24.608 "aliases": [ 00:08:24.608 "1e3d3f10-1445-432b-9e78-4e06e72648ad" 00:08:24.608 ], 00:08:24.608 "product_name": "Malloc disk", 00:08:24.608 "block_size": 512, 00:08:24.608 "num_blocks": 65536, 00:08:24.608 "uuid": "1e3d3f10-1445-432b-9e78-4e06e72648ad", 00:08:24.608 "assigned_rate_limits": { 00:08:24.608 "rw_ios_per_sec": 0, 00:08:24.608 "rw_mbytes_per_sec": 0, 00:08:24.608 "r_mbytes_per_sec": 0, 00:08:24.608 "w_mbytes_per_sec": 0 00:08:24.608 }, 00:08:24.608 "claimed": true, 00:08:24.608 "claim_type": "exclusive_write", 00:08:24.608 "zoned": false, 00:08:24.608 "supported_io_types": { 00:08:24.608 "read": true, 00:08:24.608 "write": true, 00:08:24.608 "unmap": true, 00:08:24.608 "flush": true, 00:08:24.608 "reset": true, 00:08:24.608 "nvme_admin": false, 00:08:24.608 "nvme_io": false, 00:08:24.608 "nvme_io_md": false, 00:08:24.608 "write_zeroes": true, 00:08:24.608 "zcopy": true, 00:08:24.608 "get_zone_info": false, 00:08:24.608 "zone_management": false, 00:08:24.608 "zone_append": false, 00:08:24.609 "compare": false, 00:08:24.609 "compare_and_write": false, 00:08:24.609 "abort": true, 00:08:24.609 "seek_hole": false, 00:08:24.609 "seek_data": false, 00:08:24.609 "copy": true, 00:08:24.609 "nvme_iov_md": false 00:08:24.609 }, 00:08:24.609 "memory_domains": [ 00:08:24.609 { 00:08:24.609 "dma_device_id": "system", 00:08:24.609 "dma_device_type": 1 00:08:24.609 }, 00:08:24.609 { 00:08:24.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.609 "dma_device_type": 2 00:08:24.609 } 00:08:24.609 ], 00:08:24.609 "driver_specific": {} 00:08:24.609 } 00:08:24.609 ] 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.609 "name": "Existed_Raid", 00:08:24.609 "uuid": "bc6878b5-acd6-4dec-a042-5a14590d909b", 00:08:24.609 "strip_size_kb": 64, 00:08:24.609 "state": "online", 00:08:24.609 "raid_level": "concat", 00:08:24.609 "superblock": false, 00:08:24.609 "num_base_bdevs": 3, 00:08:24.609 "num_base_bdevs_discovered": 3, 00:08:24.609 "num_base_bdevs_operational": 3, 00:08:24.609 "base_bdevs_list": [ 00:08:24.609 { 00:08:24.609 "name": "BaseBdev1", 00:08:24.609 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:24.609 "is_configured": true, 00:08:24.609 "data_offset": 0, 00:08:24.609 "data_size": 65536 00:08:24.609 }, 00:08:24.609 { 00:08:24.609 "name": "BaseBdev2", 00:08:24.609 "uuid": "843df0d6-a8b8-498e-baea-3e2a90828a89", 00:08:24.609 "is_configured": true, 00:08:24.609 "data_offset": 0, 00:08:24.609 "data_size": 65536 00:08:24.609 }, 00:08:24.609 { 00:08:24.609 "name": "BaseBdev3", 00:08:24.609 "uuid": "1e3d3f10-1445-432b-9e78-4e06e72648ad", 00:08:24.609 "is_configured": true, 00:08:24.609 "data_offset": 0, 00:08:24.609 "data_size": 65536 00:08:24.609 } 00:08:24.609 ] 00:08:24.609 }' 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.609 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:24.867 [2024-09-30 14:07:29.446643] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:24.867 "name": "Existed_Raid", 00:08:24.867 "aliases": [ 00:08:24.867 "bc6878b5-acd6-4dec-a042-5a14590d909b" 00:08:24.867 ], 00:08:24.867 "product_name": "Raid Volume", 00:08:24.867 "block_size": 512, 00:08:24.867 "num_blocks": 196608, 00:08:24.867 "uuid": "bc6878b5-acd6-4dec-a042-5a14590d909b", 00:08:24.867 "assigned_rate_limits": { 00:08:24.867 "rw_ios_per_sec": 0, 00:08:24.867 "rw_mbytes_per_sec": 0, 00:08:24.867 "r_mbytes_per_sec": 0, 00:08:24.867 "w_mbytes_per_sec": 0 00:08:24.867 }, 00:08:24.867 "claimed": false, 00:08:24.867 "zoned": false, 00:08:24.867 "supported_io_types": { 00:08:24.867 "read": true, 00:08:24.867 "write": true, 00:08:24.867 "unmap": true, 00:08:24.867 "flush": true, 00:08:24.867 "reset": true, 00:08:24.867 "nvme_admin": false, 00:08:24.867 "nvme_io": false, 00:08:24.867 "nvme_io_md": false, 00:08:24.867 "write_zeroes": true, 00:08:24.867 "zcopy": false, 00:08:24.867 "get_zone_info": false, 00:08:24.867 "zone_management": false, 00:08:24.867 "zone_append": false, 00:08:24.867 "compare": false, 00:08:24.867 "compare_and_write": false, 00:08:24.867 "abort": false, 00:08:24.867 "seek_hole": false, 00:08:24.867 "seek_data": false, 00:08:24.867 "copy": false, 00:08:24.867 "nvme_iov_md": false 00:08:24.867 }, 00:08:24.867 "memory_domains": [ 00:08:24.867 { 00:08:24.867 "dma_device_id": "system", 00:08:24.867 "dma_device_type": 1 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.867 "dma_device_type": 2 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "dma_device_id": "system", 00:08:24.867 "dma_device_type": 1 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.867 "dma_device_type": 2 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "dma_device_id": "system", 00:08:24.867 "dma_device_type": 1 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.867 "dma_device_type": 2 00:08:24.867 } 00:08:24.867 ], 00:08:24.867 "driver_specific": { 00:08:24.867 "raid": { 00:08:24.867 "uuid": "bc6878b5-acd6-4dec-a042-5a14590d909b", 00:08:24.867 "strip_size_kb": 64, 00:08:24.867 "state": "online", 00:08:24.867 "raid_level": "concat", 00:08:24.867 "superblock": false, 00:08:24.867 "num_base_bdevs": 3, 00:08:24.867 "num_base_bdevs_discovered": 3, 00:08:24.867 "num_base_bdevs_operational": 3, 00:08:24.867 "base_bdevs_list": [ 00:08:24.867 { 00:08:24.867 "name": "BaseBdev1", 00:08:24.867 "uuid": "2bfd28d1-07e6-42ef-975d-095b0c81ad29", 00:08:24.867 "is_configured": true, 00:08:24.867 "data_offset": 0, 00:08:24.867 "data_size": 65536 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "name": "BaseBdev2", 00:08:24.867 "uuid": "843df0d6-a8b8-498e-baea-3e2a90828a89", 00:08:24.867 "is_configured": true, 00:08:24.867 "data_offset": 0, 00:08:24.867 "data_size": 65536 00:08:24.867 }, 00:08:24.867 { 00:08:24.867 "name": "BaseBdev3", 00:08:24.867 "uuid": "1e3d3f10-1445-432b-9e78-4e06e72648ad", 00:08:24.867 "is_configured": true, 00:08:24.867 "data_offset": 0, 00:08:24.867 "data_size": 65536 00:08:24.867 } 00:08:24.867 ] 00:08:24.867 } 00:08:24.867 } 00:08:24.867 }' 00:08:24.867 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:25.127 BaseBdev2 00:08:25.127 BaseBdev3' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.127 [2024-09-30 14:07:29.725896] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:25.127 [2024-09-30 14:07:29.725926] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:25.127 [2024-09-30 14:07:29.725976] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.127 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.128 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.128 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.128 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.128 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.128 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.386 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.386 "name": "Existed_Raid", 00:08:25.386 "uuid": "bc6878b5-acd6-4dec-a042-5a14590d909b", 00:08:25.386 "strip_size_kb": 64, 00:08:25.386 "state": "offline", 00:08:25.386 "raid_level": "concat", 00:08:25.386 "superblock": false, 00:08:25.386 "num_base_bdevs": 3, 00:08:25.386 "num_base_bdevs_discovered": 2, 00:08:25.386 "num_base_bdevs_operational": 2, 00:08:25.386 "base_bdevs_list": [ 00:08:25.386 { 00:08:25.386 "name": null, 00:08:25.386 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.386 "is_configured": false, 00:08:25.386 "data_offset": 0, 00:08:25.386 "data_size": 65536 00:08:25.386 }, 00:08:25.386 { 00:08:25.386 "name": "BaseBdev2", 00:08:25.386 "uuid": "843df0d6-a8b8-498e-baea-3e2a90828a89", 00:08:25.386 "is_configured": true, 00:08:25.386 "data_offset": 0, 00:08:25.386 "data_size": 65536 00:08:25.386 }, 00:08:25.386 { 00:08:25.386 "name": "BaseBdev3", 00:08:25.386 "uuid": "1e3d3f10-1445-432b-9e78-4e06e72648ad", 00:08:25.386 "is_configured": true, 00:08:25.386 "data_offset": 0, 00:08:25.386 "data_size": 65536 00:08:25.386 } 00:08:25.386 ] 00:08:25.386 }' 00:08:25.386 14:07:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.386 14:07:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.645 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:25.645 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 [2024-09-30 14:07:30.152380] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 [2024-09-30 14:07:30.219287] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:25.646 [2024-09-30 14:07:30.219378] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.646 BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.646 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.905 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.905 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 [ 00:08:25.906 { 00:08:25.906 "name": "BaseBdev2", 00:08:25.906 "aliases": [ 00:08:25.906 "636b33f5-062a-4165-a9cc-83d9e655d691" 00:08:25.906 ], 00:08:25.906 "product_name": "Malloc disk", 00:08:25.906 "block_size": 512, 00:08:25.906 "num_blocks": 65536, 00:08:25.906 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:25.906 "assigned_rate_limits": { 00:08:25.906 "rw_ios_per_sec": 0, 00:08:25.906 "rw_mbytes_per_sec": 0, 00:08:25.906 "r_mbytes_per_sec": 0, 00:08:25.906 "w_mbytes_per_sec": 0 00:08:25.906 }, 00:08:25.906 "claimed": false, 00:08:25.906 "zoned": false, 00:08:25.906 "supported_io_types": { 00:08:25.906 "read": true, 00:08:25.906 "write": true, 00:08:25.906 "unmap": true, 00:08:25.906 "flush": true, 00:08:25.906 "reset": true, 00:08:25.906 "nvme_admin": false, 00:08:25.906 "nvme_io": false, 00:08:25.906 "nvme_io_md": false, 00:08:25.906 "write_zeroes": true, 00:08:25.906 "zcopy": true, 00:08:25.906 "get_zone_info": false, 00:08:25.906 "zone_management": false, 00:08:25.906 "zone_append": false, 00:08:25.906 "compare": false, 00:08:25.906 "compare_and_write": false, 00:08:25.906 "abort": true, 00:08:25.906 "seek_hole": false, 00:08:25.906 "seek_data": false, 00:08:25.906 "copy": true, 00:08:25.906 "nvme_iov_md": false 00:08:25.906 }, 00:08:25.906 "memory_domains": [ 00:08:25.906 { 00:08:25.906 "dma_device_id": "system", 00:08:25.906 "dma_device_type": 1 00:08:25.906 }, 00:08:25.906 { 00:08:25.906 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.906 "dma_device_type": 2 00:08:25.906 } 00:08:25.906 ], 00:08:25.906 "driver_specific": {} 00:08:25.906 } 00:08:25.906 ] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 BaseBdev3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 [ 00:08:25.906 { 00:08:25.906 "name": "BaseBdev3", 00:08:25.906 "aliases": [ 00:08:25.906 "b20a4607-f3af-4efb-b92e-7c0dd991231e" 00:08:25.906 ], 00:08:25.906 "product_name": "Malloc disk", 00:08:25.906 "block_size": 512, 00:08:25.906 "num_blocks": 65536, 00:08:25.906 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:25.906 "assigned_rate_limits": { 00:08:25.906 "rw_ios_per_sec": 0, 00:08:25.906 "rw_mbytes_per_sec": 0, 00:08:25.906 "r_mbytes_per_sec": 0, 00:08:25.906 "w_mbytes_per_sec": 0 00:08:25.906 }, 00:08:25.906 "claimed": false, 00:08:25.906 "zoned": false, 00:08:25.906 "supported_io_types": { 00:08:25.906 "read": true, 00:08:25.906 "write": true, 00:08:25.906 "unmap": true, 00:08:25.906 "flush": true, 00:08:25.906 "reset": true, 00:08:25.906 "nvme_admin": false, 00:08:25.906 "nvme_io": false, 00:08:25.906 "nvme_io_md": false, 00:08:25.906 "write_zeroes": true, 00:08:25.906 "zcopy": true, 00:08:25.906 "get_zone_info": false, 00:08:25.906 "zone_management": false, 00:08:25.906 "zone_append": false, 00:08:25.906 "compare": false, 00:08:25.906 "compare_and_write": false, 00:08:25.906 "abort": true, 00:08:25.906 "seek_hole": false, 00:08:25.906 "seek_data": false, 00:08:25.906 "copy": true, 00:08:25.906 "nvme_iov_md": false 00:08:25.906 }, 00:08:25.906 "memory_domains": [ 00:08:25.906 { 00:08:25.906 "dma_device_id": "system", 00:08:25.906 "dma_device_type": 1 00:08:25.906 }, 00:08:25.906 { 00:08:25.906 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.906 "dma_device_type": 2 00:08:25.906 } 00:08:25.906 ], 00:08:25.906 "driver_specific": {} 00:08:25.906 } 00:08:25.906 ] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 [2024-09-30 14:07:30.393536] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:25.906 [2024-09-30 14:07:30.393618] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:25.906 [2024-09-30 14:07:30.393653] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:25.906 [2024-09-30 14:07:30.395429] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.906 "name": "Existed_Raid", 00:08:25.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.906 "strip_size_kb": 64, 00:08:25.906 "state": "configuring", 00:08:25.906 "raid_level": "concat", 00:08:25.906 "superblock": false, 00:08:25.906 "num_base_bdevs": 3, 00:08:25.906 "num_base_bdevs_discovered": 2, 00:08:25.906 "num_base_bdevs_operational": 3, 00:08:25.906 "base_bdevs_list": [ 00:08:25.906 { 00:08:25.906 "name": "BaseBdev1", 00:08:25.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.906 "is_configured": false, 00:08:25.906 "data_offset": 0, 00:08:25.906 "data_size": 0 00:08:25.906 }, 00:08:25.906 { 00:08:25.906 "name": "BaseBdev2", 00:08:25.906 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:25.906 "is_configured": true, 00:08:25.906 "data_offset": 0, 00:08:25.906 "data_size": 65536 00:08:25.906 }, 00:08:25.906 { 00:08:25.906 "name": "BaseBdev3", 00:08:25.906 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:25.906 "is_configured": true, 00:08:25.906 "data_offset": 0, 00:08:25.906 "data_size": 65536 00:08:25.906 } 00:08:25.906 ] 00:08:25.906 }' 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.906 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.475 [2024-09-30 14:07:30.828778] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.475 "name": "Existed_Raid", 00:08:26.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.475 "strip_size_kb": 64, 00:08:26.475 "state": "configuring", 00:08:26.475 "raid_level": "concat", 00:08:26.475 "superblock": false, 00:08:26.475 "num_base_bdevs": 3, 00:08:26.475 "num_base_bdevs_discovered": 1, 00:08:26.475 "num_base_bdevs_operational": 3, 00:08:26.475 "base_bdevs_list": [ 00:08:26.475 { 00:08:26.475 "name": "BaseBdev1", 00:08:26.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.475 "is_configured": false, 00:08:26.475 "data_offset": 0, 00:08:26.475 "data_size": 0 00:08:26.475 }, 00:08:26.475 { 00:08:26.475 "name": null, 00:08:26.475 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:26.475 "is_configured": false, 00:08:26.475 "data_offset": 0, 00:08:26.475 "data_size": 65536 00:08:26.475 }, 00:08:26.475 { 00:08:26.475 "name": "BaseBdev3", 00:08:26.475 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:26.475 "is_configured": true, 00:08:26.475 "data_offset": 0, 00:08:26.475 "data_size": 65536 00:08:26.475 } 00:08:26.475 ] 00:08:26.475 }' 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.475 14:07:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.735 [2024-09-30 14:07:31.338817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:26.735 BaseBdev1 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.735 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.735 [ 00:08:26.735 { 00:08:26.736 "name": "BaseBdev1", 00:08:26.736 "aliases": [ 00:08:26.736 "e1f23c83-26ae-41b1-9d58-e90e008893d6" 00:08:26.736 ], 00:08:26.736 "product_name": "Malloc disk", 00:08:26.736 "block_size": 512, 00:08:26.736 "num_blocks": 65536, 00:08:26.736 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:26.736 "assigned_rate_limits": { 00:08:26.736 "rw_ios_per_sec": 0, 00:08:26.736 "rw_mbytes_per_sec": 0, 00:08:26.736 "r_mbytes_per_sec": 0, 00:08:26.736 "w_mbytes_per_sec": 0 00:08:26.736 }, 00:08:26.736 "claimed": true, 00:08:26.736 "claim_type": "exclusive_write", 00:08:26.736 "zoned": false, 00:08:26.736 "supported_io_types": { 00:08:26.736 "read": true, 00:08:26.736 "write": true, 00:08:26.736 "unmap": true, 00:08:26.736 "flush": true, 00:08:26.736 "reset": true, 00:08:26.736 "nvme_admin": false, 00:08:26.736 "nvme_io": false, 00:08:26.736 "nvme_io_md": false, 00:08:26.736 "write_zeroes": true, 00:08:26.736 "zcopy": true, 00:08:26.736 "get_zone_info": false, 00:08:26.736 "zone_management": false, 00:08:26.736 "zone_append": false, 00:08:26.736 "compare": false, 00:08:26.736 "compare_and_write": false, 00:08:26.736 "abort": true, 00:08:26.736 "seek_hole": false, 00:08:26.736 "seek_data": false, 00:08:26.736 "copy": true, 00:08:26.736 "nvme_iov_md": false 00:08:26.736 }, 00:08:26.736 "memory_domains": [ 00:08:26.736 { 00:08:26.736 "dma_device_id": "system", 00:08:26.736 "dma_device_type": 1 00:08:26.736 }, 00:08:26.736 { 00:08:26.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.736 "dma_device_type": 2 00:08:26.736 } 00:08:26.736 ], 00:08:26.736 "driver_specific": {} 00:08:26.736 } 00:08:26.736 ] 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.736 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.995 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.996 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.996 "name": "Existed_Raid", 00:08:26.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.996 "strip_size_kb": 64, 00:08:26.996 "state": "configuring", 00:08:26.996 "raid_level": "concat", 00:08:26.996 "superblock": false, 00:08:26.996 "num_base_bdevs": 3, 00:08:26.996 "num_base_bdevs_discovered": 2, 00:08:26.996 "num_base_bdevs_operational": 3, 00:08:26.996 "base_bdevs_list": [ 00:08:26.996 { 00:08:26.996 "name": "BaseBdev1", 00:08:26.996 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:26.996 "is_configured": true, 00:08:26.996 "data_offset": 0, 00:08:26.996 "data_size": 65536 00:08:26.996 }, 00:08:26.996 { 00:08:26.996 "name": null, 00:08:26.996 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:26.996 "is_configured": false, 00:08:26.996 "data_offset": 0, 00:08:26.996 "data_size": 65536 00:08:26.996 }, 00:08:26.996 { 00:08:26.996 "name": "BaseBdev3", 00:08:26.996 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:26.996 "is_configured": true, 00:08:26.996 "data_offset": 0, 00:08:26.996 "data_size": 65536 00:08:26.996 } 00:08:26.996 ] 00:08:26.996 }' 00:08:26.996 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.996 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.256 [2024-09-30 14:07:31.861983] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.256 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.515 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.515 "name": "Existed_Raid", 00:08:27.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.515 "strip_size_kb": 64, 00:08:27.515 "state": "configuring", 00:08:27.515 "raid_level": "concat", 00:08:27.515 "superblock": false, 00:08:27.515 "num_base_bdevs": 3, 00:08:27.515 "num_base_bdevs_discovered": 1, 00:08:27.515 "num_base_bdevs_operational": 3, 00:08:27.515 "base_bdevs_list": [ 00:08:27.515 { 00:08:27.515 "name": "BaseBdev1", 00:08:27.515 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:27.515 "is_configured": true, 00:08:27.515 "data_offset": 0, 00:08:27.515 "data_size": 65536 00:08:27.515 }, 00:08:27.515 { 00:08:27.515 "name": null, 00:08:27.515 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:27.515 "is_configured": false, 00:08:27.515 "data_offset": 0, 00:08:27.515 "data_size": 65536 00:08:27.515 }, 00:08:27.515 { 00:08:27.515 "name": null, 00:08:27.515 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:27.515 "is_configured": false, 00:08:27.515 "data_offset": 0, 00:08:27.515 "data_size": 65536 00:08:27.515 } 00:08:27.515 ] 00:08:27.515 }' 00:08:27.515 14:07:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.515 14:07:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.774 [2024-09-30 14:07:32.273289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.774 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.775 "name": "Existed_Raid", 00:08:27.775 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.775 "strip_size_kb": 64, 00:08:27.775 "state": "configuring", 00:08:27.775 "raid_level": "concat", 00:08:27.775 "superblock": false, 00:08:27.775 "num_base_bdevs": 3, 00:08:27.775 "num_base_bdevs_discovered": 2, 00:08:27.775 "num_base_bdevs_operational": 3, 00:08:27.775 "base_bdevs_list": [ 00:08:27.775 { 00:08:27.775 "name": "BaseBdev1", 00:08:27.775 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:27.775 "is_configured": true, 00:08:27.775 "data_offset": 0, 00:08:27.775 "data_size": 65536 00:08:27.775 }, 00:08:27.775 { 00:08:27.775 "name": null, 00:08:27.775 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:27.775 "is_configured": false, 00:08:27.775 "data_offset": 0, 00:08:27.775 "data_size": 65536 00:08:27.775 }, 00:08:27.775 { 00:08:27.775 "name": "BaseBdev3", 00:08:27.775 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:27.775 "is_configured": true, 00:08:27.775 "data_offset": 0, 00:08:27.775 "data_size": 65536 00:08:27.775 } 00:08:27.775 ] 00:08:27.775 }' 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.775 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.037 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.037 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:28.037 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.037 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.298 [2024-09-30 14:07:32.724568] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.298 "name": "Existed_Raid", 00:08:28.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.298 "strip_size_kb": 64, 00:08:28.298 "state": "configuring", 00:08:28.298 "raid_level": "concat", 00:08:28.298 "superblock": false, 00:08:28.298 "num_base_bdevs": 3, 00:08:28.298 "num_base_bdevs_discovered": 1, 00:08:28.298 "num_base_bdevs_operational": 3, 00:08:28.298 "base_bdevs_list": [ 00:08:28.298 { 00:08:28.298 "name": null, 00:08:28.298 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:28.298 "is_configured": false, 00:08:28.298 "data_offset": 0, 00:08:28.298 "data_size": 65536 00:08:28.298 }, 00:08:28.298 { 00:08:28.298 "name": null, 00:08:28.298 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:28.298 "is_configured": false, 00:08:28.298 "data_offset": 0, 00:08:28.298 "data_size": 65536 00:08:28.298 }, 00:08:28.298 { 00:08:28.298 "name": "BaseBdev3", 00:08:28.298 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:28.298 "is_configured": true, 00:08:28.298 "data_offset": 0, 00:08:28.298 "data_size": 65536 00:08:28.298 } 00:08:28.298 ] 00:08:28.298 }' 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.298 14:07:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.558 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.818 [2024-09-30 14:07:33.214368] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.818 "name": "Existed_Raid", 00:08:28.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.818 "strip_size_kb": 64, 00:08:28.818 "state": "configuring", 00:08:28.818 "raid_level": "concat", 00:08:28.818 "superblock": false, 00:08:28.818 "num_base_bdevs": 3, 00:08:28.818 "num_base_bdevs_discovered": 2, 00:08:28.818 "num_base_bdevs_operational": 3, 00:08:28.818 "base_bdevs_list": [ 00:08:28.818 { 00:08:28.818 "name": null, 00:08:28.818 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:28.818 "is_configured": false, 00:08:28.818 "data_offset": 0, 00:08:28.818 "data_size": 65536 00:08:28.818 }, 00:08:28.818 { 00:08:28.818 "name": "BaseBdev2", 00:08:28.818 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:28.818 "is_configured": true, 00:08:28.818 "data_offset": 0, 00:08:28.818 "data_size": 65536 00:08:28.818 }, 00:08:28.818 { 00:08:28.818 "name": "BaseBdev3", 00:08:28.818 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:28.818 "is_configured": true, 00:08:28.818 "data_offset": 0, 00:08:28.818 "data_size": 65536 00:08:28.818 } 00:08:28.818 ] 00:08:28.818 }' 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.818 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.078 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e1f23c83-26ae-41b1-9d58-e90e008893d6 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.338 [2024-09-30 14:07:33.784133] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:29.338 [2024-09-30 14:07:33.784178] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:29.338 [2024-09-30 14:07:33.784186] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:29.338 [2024-09-30 14:07:33.784441] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:08:29.338 [2024-09-30 14:07:33.784570] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:29.338 [2024-09-30 14:07:33.784587] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:08:29.338 [2024-09-30 14:07:33.784745] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.338 NewBaseBdev 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.338 [ 00:08:29.338 { 00:08:29.338 "name": "NewBaseBdev", 00:08:29.338 "aliases": [ 00:08:29.338 "e1f23c83-26ae-41b1-9d58-e90e008893d6" 00:08:29.338 ], 00:08:29.338 "product_name": "Malloc disk", 00:08:29.338 "block_size": 512, 00:08:29.338 "num_blocks": 65536, 00:08:29.338 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:29.338 "assigned_rate_limits": { 00:08:29.338 "rw_ios_per_sec": 0, 00:08:29.338 "rw_mbytes_per_sec": 0, 00:08:29.338 "r_mbytes_per_sec": 0, 00:08:29.338 "w_mbytes_per_sec": 0 00:08:29.338 }, 00:08:29.338 "claimed": true, 00:08:29.338 "claim_type": "exclusive_write", 00:08:29.338 "zoned": false, 00:08:29.338 "supported_io_types": { 00:08:29.338 "read": true, 00:08:29.338 "write": true, 00:08:29.338 "unmap": true, 00:08:29.338 "flush": true, 00:08:29.338 "reset": true, 00:08:29.338 "nvme_admin": false, 00:08:29.338 "nvme_io": false, 00:08:29.338 "nvme_io_md": false, 00:08:29.338 "write_zeroes": true, 00:08:29.338 "zcopy": true, 00:08:29.338 "get_zone_info": false, 00:08:29.338 "zone_management": false, 00:08:29.338 "zone_append": false, 00:08:29.338 "compare": false, 00:08:29.338 "compare_and_write": false, 00:08:29.338 "abort": true, 00:08:29.338 "seek_hole": false, 00:08:29.338 "seek_data": false, 00:08:29.338 "copy": true, 00:08:29.338 "nvme_iov_md": false 00:08:29.338 }, 00:08:29.338 "memory_domains": [ 00:08:29.338 { 00:08:29.338 "dma_device_id": "system", 00:08:29.338 "dma_device_type": 1 00:08:29.338 }, 00:08:29.338 { 00:08:29.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.338 "dma_device_type": 2 00:08:29.338 } 00:08:29.338 ], 00:08:29.338 "driver_specific": {} 00:08:29.338 } 00:08:29.338 ] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.338 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.339 "name": "Existed_Raid", 00:08:29.339 "uuid": "ed0f7fe2-33b3-4919-90aa-2f7e36328953", 00:08:29.339 "strip_size_kb": 64, 00:08:29.339 "state": "online", 00:08:29.339 "raid_level": "concat", 00:08:29.339 "superblock": false, 00:08:29.339 "num_base_bdevs": 3, 00:08:29.339 "num_base_bdevs_discovered": 3, 00:08:29.339 "num_base_bdevs_operational": 3, 00:08:29.339 "base_bdevs_list": [ 00:08:29.339 { 00:08:29.339 "name": "NewBaseBdev", 00:08:29.339 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:29.339 "is_configured": true, 00:08:29.339 "data_offset": 0, 00:08:29.339 "data_size": 65536 00:08:29.339 }, 00:08:29.339 { 00:08:29.339 "name": "BaseBdev2", 00:08:29.339 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:29.339 "is_configured": true, 00:08:29.339 "data_offset": 0, 00:08:29.339 "data_size": 65536 00:08:29.339 }, 00:08:29.339 { 00:08:29.339 "name": "BaseBdev3", 00:08:29.339 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:29.339 "is_configured": true, 00:08:29.339 "data_offset": 0, 00:08:29.339 "data_size": 65536 00:08:29.339 } 00:08:29.339 ] 00:08:29.339 }' 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.339 14:07:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:29.908 [2024-09-30 14:07:34.287603] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:29.908 "name": "Existed_Raid", 00:08:29.908 "aliases": [ 00:08:29.908 "ed0f7fe2-33b3-4919-90aa-2f7e36328953" 00:08:29.908 ], 00:08:29.908 "product_name": "Raid Volume", 00:08:29.908 "block_size": 512, 00:08:29.908 "num_blocks": 196608, 00:08:29.908 "uuid": "ed0f7fe2-33b3-4919-90aa-2f7e36328953", 00:08:29.908 "assigned_rate_limits": { 00:08:29.908 "rw_ios_per_sec": 0, 00:08:29.908 "rw_mbytes_per_sec": 0, 00:08:29.908 "r_mbytes_per_sec": 0, 00:08:29.908 "w_mbytes_per_sec": 0 00:08:29.908 }, 00:08:29.908 "claimed": false, 00:08:29.908 "zoned": false, 00:08:29.908 "supported_io_types": { 00:08:29.908 "read": true, 00:08:29.908 "write": true, 00:08:29.908 "unmap": true, 00:08:29.908 "flush": true, 00:08:29.908 "reset": true, 00:08:29.908 "nvme_admin": false, 00:08:29.908 "nvme_io": false, 00:08:29.908 "nvme_io_md": false, 00:08:29.908 "write_zeroes": true, 00:08:29.908 "zcopy": false, 00:08:29.908 "get_zone_info": false, 00:08:29.908 "zone_management": false, 00:08:29.908 "zone_append": false, 00:08:29.908 "compare": false, 00:08:29.908 "compare_and_write": false, 00:08:29.908 "abort": false, 00:08:29.908 "seek_hole": false, 00:08:29.908 "seek_data": false, 00:08:29.908 "copy": false, 00:08:29.908 "nvme_iov_md": false 00:08:29.908 }, 00:08:29.908 "memory_domains": [ 00:08:29.908 { 00:08:29.908 "dma_device_id": "system", 00:08:29.908 "dma_device_type": 1 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.908 "dma_device_type": 2 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "dma_device_id": "system", 00:08:29.908 "dma_device_type": 1 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.908 "dma_device_type": 2 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "dma_device_id": "system", 00:08:29.908 "dma_device_type": 1 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.908 "dma_device_type": 2 00:08:29.908 } 00:08:29.908 ], 00:08:29.908 "driver_specific": { 00:08:29.908 "raid": { 00:08:29.908 "uuid": "ed0f7fe2-33b3-4919-90aa-2f7e36328953", 00:08:29.908 "strip_size_kb": 64, 00:08:29.908 "state": "online", 00:08:29.908 "raid_level": "concat", 00:08:29.908 "superblock": false, 00:08:29.908 "num_base_bdevs": 3, 00:08:29.908 "num_base_bdevs_discovered": 3, 00:08:29.908 "num_base_bdevs_operational": 3, 00:08:29.908 "base_bdevs_list": [ 00:08:29.908 { 00:08:29.908 "name": "NewBaseBdev", 00:08:29.908 "uuid": "e1f23c83-26ae-41b1-9d58-e90e008893d6", 00:08:29.908 "is_configured": true, 00:08:29.908 "data_offset": 0, 00:08:29.908 "data_size": 65536 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "name": "BaseBdev2", 00:08:29.908 "uuid": "636b33f5-062a-4165-a9cc-83d9e655d691", 00:08:29.908 "is_configured": true, 00:08:29.908 "data_offset": 0, 00:08:29.908 "data_size": 65536 00:08:29.908 }, 00:08:29.908 { 00:08:29.908 "name": "BaseBdev3", 00:08:29.908 "uuid": "b20a4607-f3af-4efb-b92e-7c0dd991231e", 00:08:29.908 "is_configured": true, 00:08:29.908 "data_offset": 0, 00:08:29.908 "data_size": 65536 00:08:29.908 } 00:08:29.908 ] 00:08:29.908 } 00:08:29.908 } 00:08:29.908 }' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:29.908 BaseBdev2 00:08:29.908 BaseBdev3' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.908 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.909 [2024-09-30 14:07:34.550840] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:29.909 [2024-09-30 14:07:34.550866] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:29.909 [2024-09-30 14:07:34.550937] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:29.909 [2024-09-30 14:07:34.550993] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:29.909 [2024-09-30 14:07:34.551003] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 77782 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 77782 ']' 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 77782 00:08:29.909 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77782 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:30.168 killing process with pid 77782 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77782' 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 77782 00:08:30.168 [2024-09-30 14:07:34.590227] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:30.168 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 77782 00:08:30.168 [2024-09-30 14:07:34.620240] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:30.428 00:08:30.428 real 0m8.752s 00:08:30.428 user 0m14.888s 00:08:30.428 sys 0m1.821s 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.428 ************************************ 00:08:30.428 END TEST raid_state_function_test 00:08:30.428 ************************************ 00:08:30.428 14:07:34 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:08:30.428 14:07:34 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:30.428 14:07:34 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:30.428 14:07:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:30.428 ************************************ 00:08:30.428 START TEST raid_state_function_test_sb 00:08:30.428 ************************************ 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 true 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78381 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78381' 00:08:30.428 Process raid pid: 78381 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78381 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 78381 ']' 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:30.428 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:30.428 14:07:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.428 [2024-09-30 14:07:35.025121] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:30.428 [2024-09-30 14:07:35.025244] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:30.687 [2024-09-30 14:07:35.157162] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:30.687 [2024-09-30 14:07:35.185303] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:30.687 [2024-09-30 14:07:35.230690] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:30.687 [2024-09-30 14:07:35.271777] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:30.687 [2024-09-30 14:07:35.271810] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.255 [2024-09-30 14:07:35.852564] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:31.255 [2024-09-30 14:07:35.852614] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:31.255 [2024-09-30 14:07:35.852638] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:31.255 [2024-09-30 14:07:35.852647] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:31.255 [2024-09-30 14:07:35.852657] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:31.255 [2024-09-30 14:07:35.852664] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.255 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.256 "name": "Existed_Raid", 00:08:31.256 "uuid": "ac84dea2-d70e-4d03-8d8b-dac2dcadd656", 00:08:31.256 "strip_size_kb": 64, 00:08:31.256 "state": "configuring", 00:08:31.256 "raid_level": "concat", 00:08:31.256 "superblock": true, 00:08:31.256 "num_base_bdevs": 3, 00:08:31.256 "num_base_bdevs_discovered": 0, 00:08:31.256 "num_base_bdevs_operational": 3, 00:08:31.256 "base_bdevs_list": [ 00:08:31.256 { 00:08:31.256 "name": "BaseBdev1", 00:08:31.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.256 "is_configured": false, 00:08:31.256 "data_offset": 0, 00:08:31.256 "data_size": 0 00:08:31.256 }, 00:08:31.256 { 00:08:31.256 "name": "BaseBdev2", 00:08:31.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.256 "is_configured": false, 00:08:31.256 "data_offset": 0, 00:08:31.256 "data_size": 0 00:08:31.256 }, 00:08:31.256 { 00:08:31.256 "name": "BaseBdev3", 00:08:31.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.256 "is_configured": false, 00:08:31.256 "data_offset": 0, 00:08:31.256 "data_size": 0 00:08:31.256 } 00:08:31.256 ] 00:08:31.256 }' 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.256 14:07:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 [2024-09-30 14:07:36.271739] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:31.825 [2024-09-30 14:07:36.271790] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 [2024-09-30 14:07:36.279731] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:31.825 [2024-09-30 14:07:36.279771] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:31.825 [2024-09-30 14:07:36.279792] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:31.825 [2024-09-30 14:07:36.279799] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:31.825 [2024-09-30 14:07:36.279806] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:31.825 [2024-09-30 14:07:36.279813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 [2024-09-30 14:07:36.296324] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:31.825 BaseBdev1 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 [ 00:08:31.825 { 00:08:31.825 "name": "BaseBdev1", 00:08:31.825 "aliases": [ 00:08:31.825 "b8801ecd-eb9a-4bc0-aee9-9beb8f640331" 00:08:31.825 ], 00:08:31.825 "product_name": "Malloc disk", 00:08:31.825 "block_size": 512, 00:08:31.825 "num_blocks": 65536, 00:08:31.825 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:31.825 "assigned_rate_limits": { 00:08:31.825 "rw_ios_per_sec": 0, 00:08:31.825 "rw_mbytes_per_sec": 0, 00:08:31.825 "r_mbytes_per_sec": 0, 00:08:31.825 "w_mbytes_per_sec": 0 00:08:31.825 }, 00:08:31.825 "claimed": true, 00:08:31.825 "claim_type": "exclusive_write", 00:08:31.825 "zoned": false, 00:08:31.825 "supported_io_types": { 00:08:31.825 "read": true, 00:08:31.825 "write": true, 00:08:31.825 "unmap": true, 00:08:31.825 "flush": true, 00:08:31.825 "reset": true, 00:08:31.825 "nvme_admin": false, 00:08:31.825 "nvme_io": false, 00:08:31.825 "nvme_io_md": false, 00:08:31.825 "write_zeroes": true, 00:08:31.825 "zcopy": true, 00:08:31.825 "get_zone_info": false, 00:08:31.825 "zone_management": false, 00:08:31.825 "zone_append": false, 00:08:31.825 "compare": false, 00:08:31.825 "compare_and_write": false, 00:08:31.825 "abort": true, 00:08:31.825 "seek_hole": false, 00:08:31.825 "seek_data": false, 00:08:31.825 "copy": true, 00:08:31.825 "nvme_iov_md": false 00:08:31.825 }, 00:08:31.825 "memory_domains": [ 00:08:31.825 { 00:08:31.825 "dma_device_id": "system", 00:08:31.825 "dma_device_type": 1 00:08:31.825 }, 00:08:31.825 { 00:08:31.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.825 "dma_device_type": 2 00:08:31.825 } 00:08:31.825 ], 00:08:31.825 "driver_specific": {} 00:08:31.825 } 00:08:31.825 ] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.825 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.825 "name": "Existed_Raid", 00:08:31.825 "uuid": "b99a6469-13fc-4cfd-8817-a10309c7b3b4", 00:08:31.825 "strip_size_kb": 64, 00:08:31.825 "state": "configuring", 00:08:31.825 "raid_level": "concat", 00:08:31.825 "superblock": true, 00:08:31.825 "num_base_bdevs": 3, 00:08:31.825 "num_base_bdevs_discovered": 1, 00:08:31.825 "num_base_bdevs_operational": 3, 00:08:31.825 "base_bdevs_list": [ 00:08:31.825 { 00:08:31.825 "name": "BaseBdev1", 00:08:31.825 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:31.825 "is_configured": true, 00:08:31.825 "data_offset": 2048, 00:08:31.825 "data_size": 63488 00:08:31.825 }, 00:08:31.825 { 00:08:31.825 "name": "BaseBdev2", 00:08:31.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.825 "is_configured": false, 00:08:31.826 "data_offset": 0, 00:08:31.826 "data_size": 0 00:08:31.826 }, 00:08:31.826 { 00:08:31.826 "name": "BaseBdev3", 00:08:31.826 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.826 "is_configured": false, 00:08:31.826 "data_offset": 0, 00:08:31.826 "data_size": 0 00:08:31.826 } 00:08:31.826 ] 00:08:31.826 }' 00:08:31.826 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.826 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.394 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.395 [2024-09-30 14:07:36.775552] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:32.395 [2024-09-30 14:07:36.775655] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.395 [2024-09-30 14:07:36.787557] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:32.395 [2024-09-30 14:07:36.789352] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:32.395 [2024-09-30 14:07:36.789392] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:32.395 [2024-09-30 14:07:36.789404] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:32.395 [2024-09-30 14:07:36.789411] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.395 "name": "Existed_Raid", 00:08:32.395 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:32.395 "strip_size_kb": 64, 00:08:32.395 "state": "configuring", 00:08:32.395 "raid_level": "concat", 00:08:32.395 "superblock": true, 00:08:32.395 "num_base_bdevs": 3, 00:08:32.395 "num_base_bdevs_discovered": 1, 00:08:32.395 "num_base_bdevs_operational": 3, 00:08:32.395 "base_bdevs_list": [ 00:08:32.395 { 00:08:32.395 "name": "BaseBdev1", 00:08:32.395 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:32.395 "is_configured": true, 00:08:32.395 "data_offset": 2048, 00:08:32.395 "data_size": 63488 00:08:32.395 }, 00:08:32.395 { 00:08:32.395 "name": "BaseBdev2", 00:08:32.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.395 "is_configured": false, 00:08:32.395 "data_offset": 0, 00:08:32.395 "data_size": 0 00:08:32.395 }, 00:08:32.395 { 00:08:32.395 "name": "BaseBdev3", 00:08:32.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.395 "is_configured": false, 00:08:32.395 "data_offset": 0, 00:08:32.395 "data_size": 0 00:08:32.395 } 00:08:32.395 ] 00:08:32.395 }' 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.395 14:07:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.655 [2024-09-30 14:07:37.262181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:32.655 BaseBdev2 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.655 [ 00:08:32.655 { 00:08:32.655 "name": "BaseBdev2", 00:08:32.655 "aliases": [ 00:08:32.655 "b3d73f70-9dd7-4410-8a69-e11a40f4d77b" 00:08:32.655 ], 00:08:32.655 "product_name": "Malloc disk", 00:08:32.655 "block_size": 512, 00:08:32.655 "num_blocks": 65536, 00:08:32.655 "uuid": "b3d73f70-9dd7-4410-8a69-e11a40f4d77b", 00:08:32.655 "assigned_rate_limits": { 00:08:32.655 "rw_ios_per_sec": 0, 00:08:32.655 "rw_mbytes_per_sec": 0, 00:08:32.655 "r_mbytes_per_sec": 0, 00:08:32.655 "w_mbytes_per_sec": 0 00:08:32.655 }, 00:08:32.655 "claimed": true, 00:08:32.655 "claim_type": "exclusive_write", 00:08:32.655 "zoned": false, 00:08:32.655 "supported_io_types": { 00:08:32.655 "read": true, 00:08:32.655 "write": true, 00:08:32.655 "unmap": true, 00:08:32.655 "flush": true, 00:08:32.655 "reset": true, 00:08:32.655 "nvme_admin": false, 00:08:32.655 "nvme_io": false, 00:08:32.655 "nvme_io_md": false, 00:08:32.655 "write_zeroes": true, 00:08:32.655 "zcopy": true, 00:08:32.655 "get_zone_info": false, 00:08:32.655 "zone_management": false, 00:08:32.655 "zone_append": false, 00:08:32.655 "compare": false, 00:08:32.655 "compare_and_write": false, 00:08:32.655 "abort": true, 00:08:32.655 "seek_hole": false, 00:08:32.655 "seek_data": false, 00:08:32.655 "copy": true, 00:08:32.655 "nvme_iov_md": false 00:08:32.655 }, 00:08:32.655 "memory_domains": [ 00:08:32.655 { 00:08:32.655 "dma_device_id": "system", 00:08:32.655 "dma_device_type": 1 00:08:32.655 }, 00:08:32.655 { 00:08:32.655 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.655 "dma_device_type": 2 00:08:32.655 } 00:08:32.655 ], 00:08:32.655 "driver_specific": {} 00:08:32.655 } 00:08:32.655 ] 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.655 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.915 "name": "Existed_Raid", 00:08:32.915 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:32.915 "strip_size_kb": 64, 00:08:32.915 "state": "configuring", 00:08:32.915 "raid_level": "concat", 00:08:32.915 "superblock": true, 00:08:32.915 "num_base_bdevs": 3, 00:08:32.915 "num_base_bdevs_discovered": 2, 00:08:32.915 "num_base_bdevs_operational": 3, 00:08:32.915 "base_bdevs_list": [ 00:08:32.915 { 00:08:32.915 "name": "BaseBdev1", 00:08:32.915 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:32.915 "is_configured": true, 00:08:32.915 "data_offset": 2048, 00:08:32.915 "data_size": 63488 00:08:32.915 }, 00:08:32.915 { 00:08:32.915 "name": "BaseBdev2", 00:08:32.915 "uuid": "b3d73f70-9dd7-4410-8a69-e11a40f4d77b", 00:08:32.915 "is_configured": true, 00:08:32.915 "data_offset": 2048, 00:08:32.915 "data_size": 63488 00:08:32.915 }, 00:08:32.915 { 00:08:32.915 "name": "BaseBdev3", 00:08:32.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.915 "is_configured": false, 00:08:32.915 "data_offset": 0, 00:08:32.915 "data_size": 0 00:08:32.915 } 00:08:32.915 ] 00:08:32.915 }' 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.915 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.175 [2024-09-30 14:07:37.736182] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:33.175 [2024-09-30 14:07:37.736490] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:33.175 [2024-09-30 14:07:37.736541] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:33.175 BaseBdev3 00:08:33.175 [2024-09-30 14:07:37.736828] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:33.175 [2024-09-30 14:07:37.736954] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:33.175 [2024-09-30 14:07:37.736972] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:33.175 [2024-09-30 14:07:37.737087] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.175 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.175 [ 00:08:33.175 { 00:08:33.175 "name": "BaseBdev3", 00:08:33.175 "aliases": [ 00:08:33.175 "33cb1123-8d16-49a6-b726-ece395d68a81" 00:08:33.175 ], 00:08:33.175 "product_name": "Malloc disk", 00:08:33.175 "block_size": 512, 00:08:33.175 "num_blocks": 65536, 00:08:33.175 "uuid": "33cb1123-8d16-49a6-b726-ece395d68a81", 00:08:33.175 "assigned_rate_limits": { 00:08:33.175 "rw_ios_per_sec": 0, 00:08:33.175 "rw_mbytes_per_sec": 0, 00:08:33.175 "r_mbytes_per_sec": 0, 00:08:33.175 "w_mbytes_per_sec": 0 00:08:33.175 }, 00:08:33.175 "claimed": true, 00:08:33.175 "claim_type": "exclusive_write", 00:08:33.175 "zoned": false, 00:08:33.175 "supported_io_types": { 00:08:33.175 "read": true, 00:08:33.175 "write": true, 00:08:33.175 "unmap": true, 00:08:33.175 "flush": true, 00:08:33.175 "reset": true, 00:08:33.175 "nvme_admin": false, 00:08:33.175 "nvme_io": false, 00:08:33.175 "nvme_io_md": false, 00:08:33.175 "write_zeroes": true, 00:08:33.175 "zcopy": true, 00:08:33.175 "get_zone_info": false, 00:08:33.175 "zone_management": false, 00:08:33.175 "zone_append": false, 00:08:33.175 "compare": false, 00:08:33.176 "compare_and_write": false, 00:08:33.176 "abort": true, 00:08:33.176 "seek_hole": false, 00:08:33.176 "seek_data": false, 00:08:33.176 "copy": true, 00:08:33.176 "nvme_iov_md": false 00:08:33.176 }, 00:08:33.176 "memory_domains": [ 00:08:33.176 { 00:08:33.176 "dma_device_id": "system", 00:08:33.176 "dma_device_type": 1 00:08:33.176 }, 00:08:33.176 { 00:08:33.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.176 "dma_device_type": 2 00:08:33.176 } 00:08:33.176 ], 00:08:33.176 "driver_specific": {} 00:08:33.176 } 00:08:33.176 ] 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.176 "name": "Existed_Raid", 00:08:33.176 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:33.176 "strip_size_kb": 64, 00:08:33.176 "state": "online", 00:08:33.176 "raid_level": "concat", 00:08:33.176 "superblock": true, 00:08:33.176 "num_base_bdevs": 3, 00:08:33.176 "num_base_bdevs_discovered": 3, 00:08:33.176 "num_base_bdevs_operational": 3, 00:08:33.176 "base_bdevs_list": [ 00:08:33.176 { 00:08:33.176 "name": "BaseBdev1", 00:08:33.176 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:33.176 "is_configured": true, 00:08:33.176 "data_offset": 2048, 00:08:33.176 "data_size": 63488 00:08:33.176 }, 00:08:33.176 { 00:08:33.176 "name": "BaseBdev2", 00:08:33.176 "uuid": "b3d73f70-9dd7-4410-8a69-e11a40f4d77b", 00:08:33.176 "is_configured": true, 00:08:33.176 "data_offset": 2048, 00:08:33.176 "data_size": 63488 00:08:33.176 }, 00:08:33.176 { 00:08:33.176 "name": "BaseBdev3", 00:08:33.176 "uuid": "33cb1123-8d16-49a6-b726-ece395d68a81", 00:08:33.176 "is_configured": true, 00:08:33.176 "data_offset": 2048, 00:08:33.176 "data_size": 63488 00:08:33.176 } 00:08:33.176 ] 00:08:33.176 }' 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.176 14:07:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.745 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.746 [2024-09-30 14:07:38.227773] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:33.746 "name": "Existed_Raid", 00:08:33.746 "aliases": [ 00:08:33.746 "05f907ce-b693-459d-9573-7d3c7e7037d5" 00:08:33.746 ], 00:08:33.746 "product_name": "Raid Volume", 00:08:33.746 "block_size": 512, 00:08:33.746 "num_blocks": 190464, 00:08:33.746 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:33.746 "assigned_rate_limits": { 00:08:33.746 "rw_ios_per_sec": 0, 00:08:33.746 "rw_mbytes_per_sec": 0, 00:08:33.746 "r_mbytes_per_sec": 0, 00:08:33.746 "w_mbytes_per_sec": 0 00:08:33.746 }, 00:08:33.746 "claimed": false, 00:08:33.746 "zoned": false, 00:08:33.746 "supported_io_types": { 00:08:33.746 "read": true, 00:08:33.746 "write": true, 00:08:33.746 "unmap": true, 00:08:33.746 "flush": true, 00:08:33.746 "reset": true, 00:08:33.746 "nvme_admin": false, 00:08:33.746 "nvme_io": false, 00:08:33.746 "nvme_io_md": false, 00:08:33.746 "write_zeroes": true, 00:08:33.746 "zcopy": false, 00:08:33.746 "get_zone_info": false, 00:08:33.746 "zone_management": false, 00:08:33.746 "zone_append": false, 00:08:33.746 "compare": false, 00:08:33.746 "compare_and_write": false, 00:08:33.746 "abort": false, 00:08:33.746 "seek_hole": false, 00:08:33.746 "seek_data": false, 00:08:33.746 "copy": false, 00:08:33.746 "nvme_iov_md": false 00:08:33.746 }, 00:08:33.746 "memory_domains": [ 00:08:33.746 { 00:08:33.746 "dma_device_id": "system", 00:08:33.746 "dma_device_type": 1 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.746 "dma_device_type": 2 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "dma_device_id": "system", 00:08:33.746 "dma_device_type": 1 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.746 "dma_device_type": 2 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "dma_device_id": "system", 00:08:33.746 "dma_device_type": 1 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.746 "dma_device_type": 2 00:08:33.746 } 00:08:33.746 ], 00:08:33.746 "driver_specific": { 00:08:33.746 "raid": { 00:08:33.746 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:33.746 "strip_size_kb": 64, 00:08:33.746 "state": "online", 00:08:33.746 "raid_level": "concat", 00:08:33.746 "superblock": true, 00:08:33.746 "num_base_bdevs": 3, 00:08:33.746 "num_base_bdevs_discovered": 3, 00:08:33.746 "num_base_bdevs_operational": 3, 00:08:33.746 "base_bdevs_list": [ 00:08:33.746 { 00:08:33.746 "name": "BaseBdev1", 00:08:33.746 "uuid": "b8801ecd-eb9a-4bc0-aee9-9beb8f640331", 00:08:33.746 "is_configured": true, 00:08:33.746 "data_offset": 2048, 00:08:33.746 "data_size": 63488 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "name": "BaseBdev2", 00:08:33.746 "uuid": "b3d73f70-9dd7-4410-8a69-e11a40f4d77b", 00:08:33.746 "is_configured": true, 00:08:33.746 "data_offset": 2048, 00:08:33.746 "data_size": 63488 00:08:33.746 }, 00:08:33.746 { 00:08:33.746 "name": "BaseBdev3", 00:08:33.746 "uuid": "33cb1123-8d16-49a6-b726-ece395d68a81", 00:08:33.746 "is_configured": true, 00:08:33.746 "data_offset": 2048, 00:08:33.746 "data_size": 63488 00:08:33.746 } 00:08:33.746 ] 00:08:33.746 } 00:08:33.746 } 00:08:33.746 }' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:33.746 BaseBdev2 00:08:33.746 BaseBdev3' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.746 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.006 [2024-09-30 14:07:38.494954] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:34.006 [2024-09-30 14:07:38.495026] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:34.006 [2024-09-30 14:07:38.495079] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.006 "name": "Existed_Raid", 00:08:34.006 "uuid": "05f907ce-b693-459d-9573-7d3c7e7037d5", 00:08:34.006 "strip_size_kb": 64, 00:08:34.006 "state": "offline", 00:08:34.006 "raid_level": "concat", 00:08:34.006 "superblock": true, 00:08:34.006 "num_base_bdevs": 3, 00:08:34.006 "num_base_bdevs_discovered": 2, 00:08:34.006 "num_base_bdevs_operational": 2, 00:08:34.006 "base_bdevs_list": [ 00:08:34.006 { 00:08:34.006 "name": null, 00:08:34.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.006 "is_configured": false, 00:08:34.006 "data_offset": 0, 00:08:34.006 "data_size": 63488 00:08:34.006 }, 00:08:34.006 { 00:08:34.006 "name": "BaseBdev2", 00:08:34.006 "uuid": "b3d73f70-9dd7-4410-8a69-e11a40f4d77b", 00:08:34.006 "is_configured": true, 00:08:34.006 "data_offset": 2048, 00:08:34.006 "data_size": 63488 00:08:34.006 }, 00:08:34.006 { 00:08:34.006 "name": "BaseBdev3", 00:08:34.006 "uuid": "33cb1123-8d16-49a6-b726-ece395d68a81", 00:08:34.006 "is_configured": true, 00:08:34.006 "data_offset": 2048, 00:08:34.006 "data_size": 63488 00:08:34.006 } 00:08:34.006 ] 00:08:34.006 }' 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.006 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.265 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:34.265 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 [2024-09-30 14:07:38.977263] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 [2024-09-30 14:07:39.044059] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:34.524 [2024-09-30 14:07:39.044155] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 BaseBdev2 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 [ 00:08:34.524 { 00:08:34.524 "name": "BaseBdev2", 00:08:34.524 "aliases": [ 00:08:34.524 "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e" 00:08:34.524 ], 00:08:34.524 "product_name": "Malloc disk", 00:08:34.524 "block_size": 512, 00:08:34.524 "num_blocks": 65536, 00:08:34.524 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:34.524 "assigned_rate_limits": { 00:08:34.524 "rw_ios_per_sec": 0, 00:08:34.524 "rw_mbytes_per_sec": 0, 00:08:34.524 "r_mbytes_per_sec": 0, 00:08:34.524 "w_mbytes_per_sec": 0 00:08:34.524 }, 00:08:34.524 "claimed": false, 00:08:34.524 "zoned": false, 00:08:34.524 "supported_io_types": { 00:08:34.524 "read": true, 00:08:34.524 "write": true, 00:08:34.524 "unmap": true, 00:08:34.524 "flush": true, 00:08:34.524 "reset": true, 00:08:34.524 "nvme_admin": false, 00:08:34.524 "nvme_io": false, 00:08:34.524 "nvme_io_md": false, 00:08:34.524 "write_zeroes": true, 00:08:34.524 "zcopy": true, 00:08:34.524 "get_zone_info": false, 00:08:34.524 "zone_management": false, 00:08:34.524 "zone_append": false, 00:08:34.524 "compare": false, 00:08:34.524 "compare_and_write": false, 00:08:34.524 "abort": true, 00:08:34.524 "seek_hole": false, 00:08:34.524 "seek_data": false, 00:08:34.525 "copy": true, 00:08:34.525 "nvme_iov_md": false 00:08:34.525 }, 00:08:34.525 "memory_domains": [ 00:08:34.525 { 00:08:34.525 "dma_device_id": "system", 00:08:34.525 "dma_device_type": 1 00:08:34.525 }, 00:08:34.525 { 00:08:34.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.525 "dma_device_type": 2 00:08:34.525 } 00:08:34.525 ], 00:08:34.525 "driver_specific": {} 00:08:34.525 } 00:08:34.525 ] 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.525 BaseBdev3 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.525 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.784 [ 00:08:34.784 { 00:08:34.784 "name": "BaseBdev3", 00:08:34.784 "aliases": [ 00:08:34.784 "9f32877e-6bfa-43d6-90ca-a412baa84957" 00:08:34.784 ], 00:08:34.784 "product_name": "Malloc disk", 00:08:34.784 "block_size": 512, 00:08:34.784 "num_blocks": 65536, 00:08:34.784 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:34.784 "assigned_rate_limits": { 00:08:34.784 "rw_ios_per_sec": 0, 00:08:34.784 "rw_mbytes_per_sec": 0, 00:08:34.784 "r_mbytes_per_sec": 0, 00:08:34.784 "w_mbytes_per_sec": 0 00:08:34.784 }, 00:08:34.784 "claimed": false, 00:08:34.784 "zoned": false, 00:08:34.784 "supported_io_types": { 00:08:34.784 "read": true, 00:08:34.784 "write": true, 00:08:34.784 "unmap": true, 00:08:34.784 "flush": true, 00:08:34.784 "reset": true, 00:08:34.784 "nvme_admin": false, 00:08:34.784 "nvme_io": false, 00:08:34.784 "nvme_io_md": false, 00:08:34.784 "write_zeroes": true, 00:08:34.784 "zcopy": true, 00:08:34.784 "get_zone_info": false, 00:08:34.784 "zone_management": false, 00:08:34.784 "zone_append": false, 00:08:34.784 "compare": false, 00:08:34.784 "compare_and_write": false, 00:08:34.784 "abort": true, 00:08:34.784 "seek_hole": false, 00:08:34.784 "seek_data": false, 00:08:34.784 "copy": true, 00:08:34.784 "nvme_iov_md": false 00:08:34.784 }, 00:08:34.784 "memory_domains": [ 00:08:34.784 { 00:08:34.784 "dma_device_id": "system", 00:08:34.784 "dma_device_type": 1 00:08:34.784 }, 00:08:34.784 { 00:08:34.784 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.784 "dma_device_type": 2 00:08:34.784 } 00:08:34.784 ], 00:08:34.784 "driver_specific": {} 00:08:34.784 } 00:08:34.784 ] 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.784 [2024-09-30 14:07:39.218094] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:34.784 [2024-09-30 14:07:39.218140] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:34.784 [2024-09-30 14:07:39.218158] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:34.784 [2024-09-30 14:07:39.219953] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.784 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.785 "name": "Existed_Raid", 00:08:34.785 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:34.785 "strip_size_kb": 64, 00:08:34.785 "state": "configuring", 00:08:34.785 "raid_level": "concat", 00:08:34.785 "superblock": true, 00:08:34.785 "num_base_bdevs": 3, 00:08:34.785 "num_base_bdevs_discovered": 2, 00:08:34.785 "num_base_bdevs_operational": 3, 00:08:34.785 "base_bdevs_list": [ 00:08:34.785 { 00:08:34.785 "name": "BaseBdev1", 00:08:34.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.785 "is_configured": false, 00:08:34.785 "data_offset": 0, 00:08:34.785 "data_size": 0 00:08:34.785 }, 00:08:34.785 { 00:08:34.785 "name": "BaseBdev2", 00:08:34.785 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:34.785 "is_configured": true, 00:08:34.785 "data_offset": 2048, 00:08:34.785 "data_size": 63488 00:08:34.785 }, 00:08:34.785 { 00:08:34.785 "name": "BaseBdev3", 00:08:34.785 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:34.785 "is_configured": true, 00:08:34.785 "data_offset": 2048, 00:08:34.785 "data_size": 63488 00:08:34.785 } 00:08:34.785 ] 00:08:34.785 }' 00:08:34.785 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.785 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.044 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:35.044 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.044 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.044 [2024-09-30 14:07:39.669303] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:35.044 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.044 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.045 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.304 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.304 "name": "Existed_Raid", 00:08:35.304 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:35.304 "strip_size_kb": 64, 00:08:35.304 "state": "configuring", 00:08:35.304 "raid_level": "concat", 00:08:35.304 "superblock": true, 00:08:35.304 "num_base_bdevs": 3, 00:08:35.304 "num_base_bdevs_discovered": 1, 00:08:35.304 "num_base_bdevs_operational": 3, 00:08:35.304 "base_bdevs_list": [ 00:08:35.304 { 00:08:35.304 "name": "BaseBdev1", 00:08:35.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.304 "is_configured": false, 00:08:35.304 "data_offset": 0, 00:08:35.304 "data_size": 0 00:08:35.304 }, 00:08:35.304 { 00:08:35.304 "name": null, 00:08:35.304 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:35.304 "is_configured": false, 00:08:35.304 "data_offset": 0, 00:08:35.304 "data_size": 63488 00:08:35.304 }, 00:08:35.304 { 00:08:35.304 "name": "BaseBdev3", 00:08:35.304 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:35.304 "is_configured": true, 00:08:35.304 "data_offset": 2048, 00:08:35.304 "data_size": 63488 00:08:35.304 } 00:08:35.304 ] 00:08:35.304 }' 00:08:35.304 14:07:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.304 14:07:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 [2024-09-30 14:07:40.119279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:35.564 BaseBdev1 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 [ 00:08:35.564 { 00:08:35.564 "name": "BaseBdev1", 00:08:35.564 "aliases": [ 00:08:35.564 "4964d294-98ef-45a2-b9dc-59873298ca03" 00:08:35.564 ], 00:08:35.564 "product_name": "Malloc disk", 00:08:35.564 "block_size": 512, 00:08:35.564 "num_blocks": 65536, 00:08:35.564 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:35.564 "assigned_rate_limits": { 00:08:35.564 "rw_ios_per_sec": 0, 00:08:35.564 "rw_mbytes_per_sec": 0, 00:08:35.564 "r_mbytes_per_sec": 0, 00:08:35.564 "w_mbytes_per_sec": 0 00:08:35.564 }, 00:08:35.564 "claimed": true, 00:08:35.564 "claim_type": "exclusive_write", 00:08:35.564 "zoned": false, 00:08:35.564 "supported_io_types": { 00:08:35.564 "read": true, 00:08:35.564 "write": true, 00:08:35.564 "unmap": true, 00:08:35.564 "flush": true, 00:08:35.564 "reset": true, 00:08:35.564 "nvme_admin": false, 00:08:35.564 "nvme_io": false, 00:08:35.564 "nvme_io_md": false, 00:08:35.564 "write_zeroes": true, 00:08:35.564 "zcopy": true, 00:08:35.564 "get_zone_info": false, 00:08:35.564 "zone_management": false, 00:08:35.564 "zone_append": false, 00:08:35.564 "compare": false, 00:08:35.564 "compare_and_write": false, 00:08:35.564 "abort": true, 00:08:35.564 "seek_hole": false, 00:08:35.564 "seek_data": false, 00:08:35.564 "copy": true, 00:08:35.564 "nvme_iov_md": false 00:08:35.564 }, 00:08:35.564 "memory_domains": [ 00:08:35.564 { 00:08:35.564 "dma_device_id": "system", 00:08:35.564 "dma_device_type": 1 00:08:35.564 }, 00:08:35.564 { 00:08:35.564 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:35.564 "dma_device_type": 2 00:08:35.564 } 00:08:35.564 ], 00:08:35.564 "driver_specific": {} 00:08:35.564 } 00:08:35.564 ] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.564 "name": "Existed_Raid", 00:08:35.564 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:35.564 "strip_size_kb": 64, 00:08:35.564 "state": "configuring", 00:08:35.564 "raid_level": "concat", 00:08:35.564 "superblock": true, 00:08:35.564 "num_base_bdevs": 3, 00:08:35.564 "num_base_bdevs_discovered": 2, 00:08:35.564 "num_base_bdevs_operational": 3, 00:08:35.564 "base_bdevs_list": [ 00:08:35.564 { 00:08:35.564 "name": "BaseBdev1", 00:08:35.564 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:35.564 "is_configured": true, 00:08:35.564 "data_offset": 2048, 00:08:35.564 "data_size": 63488 00:08:35.564 }, 00:08:35.564 { 00:08:35.564 "name": null, 00:08:35.564 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:35.564 "is_configured": false, 00:08:35.564 "data_offset": 0, 00:08:35.564 "data_size": 63488 00:08:35.564 }, 00:08:35.564 { 00:08:35.564 "name": "BaseBdev3", 00:08:35.564 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:35.564 "is_configured": true, 00:08:35.564 "data_offset": 2048, 00:08:35.564 "data_size": 63488 00:08:35.564 } 00:08:35.564 ] 00:08:35.564 }' 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.564 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.133 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:36.133 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.133 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.133 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.134 [2024-09-30 14:07:40.654522] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.134 "name": "Existed_Raid", 00:08:36.134 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:36.134 "strip_size_kb": 64, 00:08:36.134 "state": "configuring", 00:08:36.134 "raid_level": "concat", 00:08:36.134 "superblock": true, 00:08:36.134 "num_base_bdevs": 3, 00:08:36.134 "num_base_bdevs_discovered": 1, 00:08:36.134 "num_base_bdevs_operational": 3, 00:08:36.134 "base_bdevs_list": [ 00:08:36.134 { 00:08:36.134 "name": "BaseBdev1", 00:08:36.134 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:36.134 "is_configured": true, 00:08:36.134 "data_offset": 2048, 00:08:36.134 "data_size": 63488 00:08:36.134 }, 00:08:36.134 { 00:08:36.134 "name": null, 00:08:36.134 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:36.134 "is_configured": false, 00:08:36.134 "data_offset": 0, 00:08:36.134 "data_size": 63488 00:08:36.134 }, 00:08:36.134 { 00:08:36.134 "name": null, 00:08:36.134 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:36.134 "is_configured": false, 00:08:36.134 "data_offset": 0, 00:08:36.134 "data_size": 63488 00:08:36.134 } 00:08:36.134 ] 00:08:36.134 }' 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.134 14:07:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 [2024-09-30 14:07:41.197584] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.703 "name": "Existed_Raid", 00:08:36.703 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:36.703 "strip_size_kb": 64, 00:08:36.703 "state": "configuring", 00:08:36.703 "raid_level": "concat", 00:08:36.703 "superblock": true, 00:08:36.703 "num_base_bdevs": 3, 00:08:36.703 "num_base_bdevs_discovered": 2, 00:08:36.703 "num_base_bdevs_operational": 3, 00:08:36.703 "base_bdevs_list": [ 00:08:36.703 { 00:08:36.703 "name": "BaseBdev1", 00:08:36.703 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:36.703 "is_configured": true, 00:08:36.703 "data_offset": 2048, 00:08:36.703 "data_size": 63488 00:08:36.703 }, 00:08:36.703 { 00:08:36.703 "name": null, 00:08:36.703 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:36.703 "is_configured": false, 00:08:36.703 "data_offset": 0, 00:08:36.703 "data_size": 63488 00:08:36.703 }, 00:08:36.703 { 00:08:36.703 "name": "BaseBdev3", 00:08:36.703 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:36.703 "is_configured": true, 00:08:36.703 "data_offset": 2048, 00:08:36.703 "data_size": 63488 00:08:36.703 } 00:08:36.703 ] 00:08:36.703 }' 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.703 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.271 [2024-09-30 14:07:41.680766] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.271 "name": "Existed_Raid", 00:08:37.271 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:37.271 "strip_size_kb": 64, 00:08:37.271 "state": "configuring", 00:08:37.271 "raid_level": "concat", 00:08:37.271 "superblock": true, 00:08:37.271 "num_base_bdevs": 3, 00:08:37.271 "num_base_bdevs_discovered": 1, 00:08:37.271 "num_base_bdevs_operational": 3, 00:08:37.271 "base_bdevs_list": [ 00:08:37.271 { 00:08:37.271 "name": null, 00:08:37.271 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:37.271 "is_configured": false, 00:08:37.271 "data_offset": 0, 00:08:37.271 "data_size": 63488 00:08:37.271 }, 00:08:37.271 { 00:08:37.271 "name": null, 00:08:37.271 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:37.271 "is_configured": false, 00:08:37.271 "data_offset": 0, 00:08:37.271 "data_size": 63488 00:08:37.271 }, 00:08:37.271 { 00:08:37.271 "name": "BaseBdev3", 00:08:37.271 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:37.271 "is_configured": true, 00:08:37.271 "data_offset": 2048, 00:08:37.271 "data_size": 63488 00:08:37.271 } 00:08:37.271 ] 00:08:37.271 }' 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.271 14:07:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.531 [2024-09-30 14:07:42.154258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.531 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.790 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.790 "name": "Existed_Raid", 00:08:37.790 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:37.790 "strip_size_kb": 64, 00:08:37.790 "state": "configuring", 00:08:37.790 "raid_level": "concat", 00:08:37.790 "superblock": true, 00:08:37.790 "num_base_bdevs": 3, 00:08:37.790 "num_base_bdevs_discovered": 2, 00:08:37.790 "num_base_bdevs_operational": 3, 00:08:37.790 "base_bdevs_list": [ 00:08:37.790 { 00:08:37.790 "name": null, 00:08:37.790 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:37.790 "is_configured": false, 00:08:37.790 "data_offset": 0, 00:08:37.790 "data_size": 63488 00:08:37.790 }, 00:08:37.790 { 00:08:37.790 "name": "BaseBdev2", 00:08:37.790 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:37.790 "is_configured": true, 00:08:37.790 "data_offset": 2048, 00:08:37.790 "data_size": 63488 00:08:37.790 }, 00:08:37.790 { 00:08:37.790 "name": "BaseBdev3", 00:08:37.790 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:37.790 "is_configured": true, 00:08:37.790 "data_offset": 2048, 00:08:37.790 "data_size": 63488 00:08:37.790 } 00:08:37.790 ] 00:08:37.790 }' 00:08:37.790 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.790 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 4964d294-98ef-45a2-b9dc-59873298ca03 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 [2024-09-30 14:07:42.672354] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:38.050 [2024-09-30 14:07:42.672526] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:38.050 [2024-09-30 14:07:42.672546] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:38.050 [2024-09-30 14:07:42.672785] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:08:38.050 [2024-09-30 14:07:42.672892] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:38.050 [2024-09-30 14:07:42.672911] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:08:38.050 [2024-09-30 14:07:42.673009] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:38.050 NewBaseBdev 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.050 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.050 [ 00:08:38.050 { 00:08:38.050 "name": "NewBaseBdev", 00:08:38.050 "aliases": [ 00:08:38.050 "4964d294-98ef-45a2-b9dc-59873298ca03" 00:08:38.050 ], 00:08:38.050 "product_name": "Malloc disk", 00:08:38.050 "block_size": 512, 00:08:38.050 "num_blocks": 65536, 00:08:38.051 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:38.051 "assigned_rate_limits": { 00:08:38.051 "rw_ios_per_sec": 0, 00:08:38.051 "rw_mbytes_per_sec": 0, 00:08:38.051 "r_mbytes_per_sec": 0, 00:08:38.051 "w_mbytes_per_sec": 0 00:08:38.051 }, 00:08:38.051 "claimed": true, 00:08:38.051 "claim_type": "exclusive_write", 00:08:38.051 "zoned": false, 00:08:38.051 "supported_io_types": { 00:08:38.051 "read": true, 00:08:38.051 "write": true, 00:08:38.051 "unmap": true, 00:08:38.051 "flush": true, 00:08:38.051 "reset": true, 00:08:38.051 "nvme_admin": false, 00:08:38.051 "nvme_io": false, 00:08:38.051 "nvme_io_md": false, 00:08:38.051 "write_zeroes": true, 00:08:38.051 "zcopy": true, 00:08:38.051 "get_zone_info": false, 00:08:38.051 "zone_management": false, 00:08:38.051 "zone_append": false, 00:08:38.051 "compare": false, 00:08:38.051 "compare_and_write": false, 00:08:38.051 "abort": true, 00:08:38.051 "seek_hole": false, 00:08:38.051 "seek_data": false, 00:08:38.310 "copy": true, 00:08:38.310 "nvme_iov_md": false 00:08:38.310 }, 00:08:38.310 "memory_domains": [ 00:08:38.310 { 00:08:38.310 "dma_device_id": "system", 00:08:38.310 "dma_device_type": 1 00:08:38.310 }, 00:08:38.310 { 00:08:38.310 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.310 "dma_device_type": 2 00:08:38.310 } 00:08:38.310 ], 00:08:38.310 "driver_specific": {} 00:08:38.310 } 00:08:38.310 ] 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.310 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.310 "name": "Existed_Raid", 00:08:38.310 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:38.310 "strip_size_kb": 64, 00:08:38.310 "state": "online", 00:08:38.310 "raid_level": "concat", 00:08:38.310 "superblock": true, 00:08:38.310 "num_base_bdevs": 3, 00:08:38.310 "num_base_bdevs_discovered": 3, 00:08:38.310 "num_base_bdevs_operational": 3, 00:08:38.310 "base_bdevs_list": [ 00:08:38.310 { 00:08:38.310 "name": "NewBaseBdev", 00:08:38.310 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:38.310 "is_configured": true, 00:08:38.310 "data_offset": 2048, 00:08:38.310 "data_size": 63488 00:08:38.310 }, 00:08:38.310 { 00:08:38.310 "name": "BaseBdev2", 00:08:38.310 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:38.310 "is_configured": true, 00:08:38.310 "data_offset": 2048, 00:08:38.310 "data_size": 63488 00:08:38.310 }, 00:08:38.310 { 00:08:38.310 "name": "BaseBdev3", 00:08:38.311 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:38.311 "is_configured": true, 00:08:38.311 "data_offset": 2048, 00:08:38.311 "data_size": 63488 00:08:38.311 } 00:08:38.311 ] 00:08:38.311 }' 00:08:38.311 14:07:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.311 14:07:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.570 [2024-09-30 14:07:43.131899] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:38.570 "name": "Existed_Raid", 00:08:38.570 "aliases": [ 00:08:38.570 "032b0994-e047-4229-8fa8-a76ff3c60fb8" 00:08:38.570 ], 00:08:38.570 "product_name": "Raid Volume", 00:08:38.570 "block_size": 512, 00:08:38.570 "num_blocks": 190464, 00:08:38.570 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:38.570 "assigned_rate_limits": { 00:08:38.570 "rw_ios_per_sec": 0, 00:08:38.570 "rw_mbytes_per_sec": 0, 00:08:38.570 "r_mbytes_per_sec": 0, 00:08:38.570 "w_mbytes_per_sec": 0 00:08:38.570 }, 00:08:38.570 "claimed": false, 00:08:38.570 "zoned": false, 00:08:38.570 "supported_io_types": { 00:08:38.570 "read": true, 00:08:38.570 "write": true, 00:08:38.570 "unmap": true, 00:08:38.570 "flush": true, 00:08:38.570 "reset": true, 00:08:38.570 "nvme_admin": false, 00:08:38.570 "nvme_io": false, 00:08:38.570 "nvme_io_md": false, 00:08:38.570 "write_zeroes": true, 00:08:38.570 "zcopy": false, 00:08:38.570 "get_zone_info": false, 00:08:38.570 "zone_management": false, 00:08:38.570 "zone_append": false, 00:08:38.570 "compare": false, 00:08:38.570 "compare_and_write": false, 00:08:38.570 "abort": false, 00:08:38.570 "seek_hole": false, 00:08:38.570 "seek_data": false, 00:08:38.570 "copy": false, 00:08:38.570 "nvme_iov_md": false 00:08:38.570 }, 00:08:38.570 "memory_domains": [ 00:08:38.570 { 00:08:38.570 "dma_device_id": "system", 00:08:38.570 "dma_device_type": 1 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.570 "dma_device_type": 2 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "dma_device_id": "system", 00:08:38.570 "dma_device_type": 1 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.570 "dma_device_type": 2 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "dma_device_id": "system", 00:08:38.570 "dma_device_type": 1 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.570 "dma_device_type": 2 00:08:38.570 } 00:08:38.570 ], 00:08:38.570 "driver_specific": { 00:08:38.570 "raid": { 00:08:38.570 "uuid": "032b0994-e047-4229-8fa8-a76ff3c60fb8", 00:08:38.570 "strip_size_kb": 64, 00:08:38.570 "state": "online", 00:08:38.570 "raid_level": "concat", 00:08:38.570 "superblock": true, 00:08:38.570 "num_base_bdevs": 3, 00:08:38.570 "num_base_bdevs_discovered": 3, 00:08:38.570 "num_base_bdevs_operational": 3, 00:08:38.570 "base_bdevs_list": [ 00:08:38.570 { 00:08:38.570 "name": "NewBaseBdev", 00:08:38.570 "uuid": "4964d294-98ef-45a2-b9dc-59873298ca03", 00:08:38.570 "is_configured": true, 00:08:38.570 "data_offset": 2048, 00:08:38.570 "data_size": 63488 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "name": "BaseBdev2", 00:08:38.570 "uuid": "8b803ccb-4ab7-4c15-9a23-4075f84f6f9e", 00:08:38.570 "is_configured": true, 00:08:38.570 "data_offset": 2048, 00:08:38.570 "data_size": 63488 00:08:38.570 }, 00:08:38.570 { 00:08:38.570 "name": "BaseBdev3", 00:08:38.570 "uuid": "9f32877e-6bfa-43d6-90ca-a412baa84957", 00:08:38.570 "is_configured": true, 00:08:38.570 "data_offset": 2048, 00:08:38.570 "data_size": 63488 00:08:38.570 } 00:08:38.570 ] 00:08:38.570 } 00:08:38.570 } 00:08:38.570 }' 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:38.570 BaseBdev2 00:08:38.570 BaseBdev3' 00:08:38.570 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.830 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.831 [2024-09-30 14:07:43.415112] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:38.831 [2024-09-30 14:07:43.415141] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:38.831 [2024-09-30 14:07:43.415200] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:38.831 [2024-09-30 14:07:43.415251] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:38.831 [2024-09-30 14:07:43.415261] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78381 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 78381 ']' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 78381 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78381 00:08:38.831 killing process with pid 78381 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78381' 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 78381 00:08:38.831 [2024-09-30 14:07:43.450237] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:38.831 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 78381 00:08:38.831 [2024-09-30 14:07:43.480119] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:39.090 14:07:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:39.090 00:08:39.090 real 0m8.789s 00:08:39.090 user 0m15.025s 00:08:39.090 sys 0m1.794s 00:08:39.090 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:39.090 14:07:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.090 ************************************ 00:08:39.090 END TEST raid_state_function_test_sb 00:08:39.090 ************************************ 00:08:39.350 14:07:43 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:39.350 14:07:43 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:39.350 14:07:43 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:39.350 14:07:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:39.350 ************************************ 00:08:39.350 START TEST raid_superblock_test 00:08:39.350 ************************************ 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 3 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=78986 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 78986 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 78986 ']' 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:39.350 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:39.350 14:07:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.350 [2024-09-30 14:07:43.879586] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:39.350 [2024-09-30 14:07:43.879720] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78986 ] 00:08:39.609 [2024-09-30 14:07:44.010349] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:39.609 [2024-09-30 14:07:44.039830] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.609 [2024-09-30 14:07:44.085071] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.609 [2024-09-30 14:07:44.126012] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.609 [2024-09-30 14:07:44.126052] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 malloc1 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 [2024-09-30 14:07:44.731442] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:40.179 [2024-09-30 14:07:44.731520] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.179 [2024-09-30 14:07:44.731544] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:40.179 [2024-09-30 14:07:44.731555] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.179 [2024-09-30 14:07:44.733526] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.179 [2024-09-30 14:07:44.733555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:40.179 pt1 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 malloc2 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 [2024-09-30 14:07:44.769824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:40.179 [2024-09-30 14:07:44.769882] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.179 [2024-09-30 14:07:44.769904] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:40.179 [2024-09-30 14:07:44.769915] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.179 [2024-09-30 14:07:44.772435] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.179 [2024-09-30 14:07:44.772470] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:40.179 pt2 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 malloc3 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 [2024-09-30 14:07:44.798171] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:40.179 [2024-09-30 14:07:44.798216] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.179 [2024-09-30 14:07:44.798234] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:40.179 [2024-09-30 14:07:44.798242] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.179 [2024-09-30 14:07:44.800222] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.179 [2024-09-30 14:07:44.800256] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:40.179 pt3 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.179 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 [2024-09-30 14:07:44.810219] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:40.179 [2024-09-30 14:07:44.812005] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:40.179 [2024-09-30 14:07:44.812083] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:40.179 [2024-09-30 14:07:44.812215] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:40.179 [2024-09-30 14:07:44.812228] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:40.179 [2024-09-30 14:07:44.812455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:40.179 [2024-09-30 14:07:44.812590] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:40.179 [2024-09-30 14:07:44.812611] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:08:40.180 [2024-09-30 14:07:44.812727] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.180 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:40.438 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.438 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.438 "name": "raid_bdev1", 00:08:40.438 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:40.438 "strip_size_kb": 64, 00:08:40.438 "state": "online", 00:08:40.438 "raid_level": "concat", 00:08:40.438 "superblock": true, 00:08:40.438 "num_base_bdevs": 3, 00:08:40.438 "num_base_bdevs_discovered": 3, 00:08:40.438 "num_base_bdevs_operational": 3, 00:08:40.438 "base_bdevs_list": [ 00:08:40.438 { 00:08:40.438 "name": "pt1", 00:08:40.438 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:40.438 "is_configured": true, 00:08:40.438 "data_offset": 2048, 00:08:40.438 "data_size": 63488 00:08:40.438 }, 00:08:40.438 { 00:08:40.438 "name": "pt2", 00:08:40.438 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:40.438 "is_configured": true, 00:08:40.438 "data_offset": 2048, 00:08:40.438 "data_size": 63488 00:08:40.438 }, 00:08:40.438 { 00:08:40.438 "name": "pt3", 00:08:40.438 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:40.438 "is_configured": true, 00:08:40.438 "data_offset": 2048, 00:08:40.438 "data_size": 63488 00:08:40.438 } 00:08:40.438 ] 00:08:40.438 }' 00:08:40.438 14:07:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.438 14:07:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.696 [2024-09-30 14:07:45.289660] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:40.696 "name": "raid_bdev1", 00:08:40.696 "aliases": [ 00:08:40.696 "6a3abf78-bbb1-4511-8b96-1d6840448253" 00:08:40.696 ], 00:08:40.696 "product_name": "Raid Volume", 00:08:40.696 "block_size": 512, 00:08:40.696 "num_blocks": 190464, 00:08:40.696 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:40.696 "assigned_rate_limits": { 00:08:40.696 "rw_ios_per_sec": 0, 00:08:40.696 "rw_mbytes_per_sec": 0, 00:08:40.696 "r_mbytes_per_sec": 0, 00:08:40.696 "w_mbytes_per_sec": 0 00:08:40.696 }, 00:08:40.696 "claimed": false, 00:08:40.696 "zoned": false, 00:08:40.696 "supported_io_types": { 00:08:40.696 "read": true, 00:08:40.696 "write": true, 00:08:40.696 "unmap": true, 00:08:40.696 "flush": true, 00:08:40.696 "reset": true, 00:08:40.696 "nvme_admin": false, 00:08:40.696 "nvme_io": false, 00:08:40.696 "nvme_io_md": false, 00:08:40.696 "write_zeroes": true, 00:08:40.696 "zcopy": false, 00:08:40.696 "get_zone_info": false, 00:08:40.696 "zone_management": false, 00:08:40.696 "zone_append": false, 00:08:40.696 "compare": false, 00:08:40.696 "compare_and_write": false, 00:08:40.696 "abort": false, 00:08:40.696 "seek_hole": false, 00:08:40.696 "seek_data": false, 00:08:40.696 "copy": false, 00:08:40.696 "nvme_iov_md": false 00:08:40.696 }, 00:08:40.696 "memory_domains": [ 00:08:40.696 { 00:08:40.696 "dma_device_id": "system", 00:08:40.696 "dma_device_type": 1 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.696 "dma_device_type": 2 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "dma_device_id": "system", 00:08:40.696 "dma_device_type": 1 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.696 "dma_device_type": 2 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "dma_device_id": "system", 00:08:40.696 "dma_device_type": 1 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.696 "dma_device_type": 2 00:08:40.696 } 00:08:40.696 ], 00:08:40.696 "driver_specific": { 00:08:40.696 "raid": { 00:08:40.696 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:40.696 "strip_size_kb": 64, 00:08:40.696 "state": "online", 00:08:40.696 "raid_level": "concat", 00:08:40.696 "superblock": true, 00:08:40.696 "num_base_bdevs": 3, 00:08:40.696 "num_base_bdevs_discovered": 3, 00:08:40.696 "num_base_bdevs_operational": 3, 00:08:40.696 "base_bdevs_list": [ 00:08:40.696 { 00:08:40.696 "name": "pt1", 00:08:40.696 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:40.696 "is_configured": true, 00:08:40.696 "data_offset": 2048, 00:08:40.696 "data_size": 63488 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "name": "pt2", 00:08:40.696 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:40.696 "is_configured": true, 00:08:40.696 "data_offset": 2048, 00:08:40.696 "data_size": 63488 00:08:40.696 }, 00:08:40.696 { 00:08:40.696 "name": "pt3", 00:08:40.696 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:40.696 "is_configured": true, 00:08:40.696 "data_offset": 2048, 00:08:40.696 "data_size": 63488 00:08:40.696 } 00:08:40.696 ] 00:08:40.696 } 00:08:40.696 } 00:08:40.696 }' 00:08:40.696 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:40.955 pt2 00:08:40.955 pt3' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.955 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.956 [2024-09-30 14:07:45.553111] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=6a3abf78-bbb1-4511-8b96-1d6840448253 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 6a3abf78-bbb1-4511-8b96-1d6840448253 ']' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.956 [2024-09-30 14:07:45.596796] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:40.956 [2024-09-30 14:07:45.596863] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:40.956 [2024-09-30 14:07:45.596954] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:40.956 [2024-09-30 14:07:45.597036] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:40.956 [2024-09-30 14:07:45.597070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.956 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.214 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.214 [2024-09-30 14:07:45.764591] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:41.214 [2024-09-30 14:07:45.766448] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:41.214 [2024-09-30 14:07:45.766554] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:41.214 [2024-09-30 14:07:45.766627] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:41.214 [2024-09-30 14:07:45.766721] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:41.215 [2024-09-30 14:07:45.766784] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:41.215 [2024-09-30 14:07:45.766855] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:41.215 [2024-09-30 14:07:45.766888] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:08:41.215 request: 00:08:41.215 { 00:08:41.215 "name": "raid_bdev1", 00:08:41.215 "raid_level": "concat", 00:08:41.215 "base_bdevs": [ 00:08:41.215 "malloc1", 00:08:41.215 "malloc2", 00:08:41.215 "malloc3" 00:08:41.215 ], 00:08:41.215 "strip_size_kb": 64, 00:08:41.215 "superblock": false, 00:08:41.215 "method": "bdev_raid_create", 00:08:41.215 "req_id": 1 00:08:41.215 } 00:08:41.215 Got JSON-RPC error response 00:08:41.215 response: 00:08:41.215 { 00:08:41.215 "code": -17, 00:08:41.215 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:41.215 } 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.215 [2024-09-30 14:07:45.828436] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:41.215 [2024-09-30 14:07:45.828532] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:41.215 [2024-09-30 14:07:45.828567] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:41.215 [2024-09-30 14:07:45.828593] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:41.215 [2024-09-30 14:07:45.830589] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:41.215 [2024-09-30 14:07:45.830677] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:41.215 [2024-09-30 14:07:45.830760] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:41.215 [2024-09-30 14:07:45.830825] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:41.215 pt1 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:41.215 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.473 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.473 "name": "raid_bdev1", 00:08:41.473 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:41.473 "strip_size_kb": 64, 00:08:41.473 "state": "configuring", 00:08:41.473 "raid_level": "concat", 00:08:41.473 "superblock": true, 00:08:41.473 "num_base_bdevs": 3, 00:08:41.473 "num_base_bdevs_discovered": 1, 00:08:41.473 "num_base_bdevs_operational": 3, 00:08:41.473 "base_bdevs_list": [ 00:08:41.473 { 00:08:41.473 "name": "pt1", 00:08:41.473 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:41.473 "is_configured": true, 00:08:41.473 "data_offset": 2048, 00:08:41.473 "data_size": 63488 00:08:41.473 }, 00:08:41.473 { 00:08:41.473 "name": null, 00:08:41.473 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:41.473 "is_configured": false, 00:08:41.473 "data_offset": 2048, 00:08:41.473 "data_size": 63488 00:08:41.473 }, 00:08:41.473 { 00:08:41.473 "name": null, 00:08:41.474 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:41.474 "is_configured": false, 00:08:41.474 "data_offset": 2048, 00:08:41.474 "data_size": 63488 00:08:41.474 } 00:08:41.474 ] 00:08:41.474 }' 00:08:41.474 14:07:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.474 14:07:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.732 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:41.732 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:41.732 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.732 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.732 [2024-09-30 14:07:46.267705] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:41.732 [2024-09-30 14:07:46.267820] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:41.732 [2024-09-30 14:07:46.267849] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:08:41.732 [2024-09-30 14:07:46.267858] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:41.732 [2024-09-30 14:07:46.268224] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:41.732 [2024-09-30 14:07:46.268241] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:41.733 [2024-09-30 14:07:46.268311] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:41.733 [2024-09-30 14:07:46.268330] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:41.733 pt2 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.733 [2024-09-30 14:07:46.279706] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.733 "name": "raid_bdev1", 00:08:41.733 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:41.733 "strip_size_kb": 64, 00:08:41.733 "state": "configuring", 00:08:41.733 "raid_level": "concat", 00:08:41.733 "superblock": true, 00:08:41.733 "num_base_bdevs": 3, 00:08:41.733 "num_base_bdevs_discovered": 1, 00:08:41.733 "num_base_bdevs_operational": 3, 00:08:41.733 "base_bdevs_list": [ 00:08:41.733 { 00:08:41.733 "name": "pt1", 00:08:41.733 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:41.733 "is_configured": true, 00:08:41.733 "data_offset": 2048, 00:08:41.733 "data_size": 63488 00:08:41.733 }, 00:08:41.733 { 00:08:41.733 "name": null, 00:08:41.733 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:41.733 "is_configured": false, 00:08:41.733 "data_offset": 0, 00:08:41.733 "data_size": 63488 00:08:41.733 }, 00:08:41.733 { 00:08:41.733 "name": null, 00:08:41.733 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:41.733 "is_configured": false, 00:08:41.733 "data_offset": 2048, 00:08:41.733 "data_size": 63488 00:08:41.733 } 00:08:41.733 ] 00:08:41.733 }' 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.733 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.310 [2024-09-30 14:07:46.738896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:42.310 [2024-09-30 14:07:46.739030] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:42.310 [2024-09-30 14:07:46.739064] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:42.310 [2024-09-30 14:07:46.739094] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:42.310 [2024-09-30 14:07:46.739548] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:42.310 [2024-09-30 14:07:46.739610] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:42.310 [2024-09-30 14:07:46.739713] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:42.310 [2024-09-30 14:07:46.739766] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:42.310 pt2 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.310 [2024-09-30 14:07:46.750864] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:42.310 [2024-09-30 14:07:46.750950] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:42.310 [2024-09-30 14:07:46.750977] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:42.310 [2024-09-30 14:07:46.751006] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:42.310 [2024-09-30 14:07:46.751325] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:42.310 [2024-09-30 14:07:46.751380] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:42.310 [2024-09-30 14:07:46.751460] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:42.310 [2024-09-30 14:07:46.751525] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:42.310 [2024-09-30 14:07:46.751638] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:42.310 [2024-09-30 14:07:46.751676] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:42.310 [2024-09-30 14:07:46.751903] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:08:42.310 [2024-09-30 14:07:46.752035] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:42.310 [2024-09-30 14:07:46.752067] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:08:42.310 [2024-09-30 14:07:46.752191] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:42.310 pt3 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.310 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.310 "name": "raid_bdev1", 00:08:42.310 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:42.310 "strip_size_kb": 64, 00:08:42.310 "state": "online", 00:08:42.310 "raid_level": "concat", 00:08:42.310 "superblock": true, 00:08:42.310 "num_base_bdevs": 3, 00:08:42.310 "num_base_bdevs_discovered": 3, 00:08:42.310 "num_base_bdevs_operational": 3, 00:08:42.310 "base_bdevs_list": [ 00:08:42.310 { 00:08:42.310 "name": "pt1", 00:08:42.310 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:42.310 "is_configured": true, 00:08:42.310 "data_offset": 2048, 00:08:42.310 "data_size": 63488 00:08:42.310 }, 00:08:42.311 { 00:08:42.311 "name": "pt2", 00:08:42.311 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:42.311 "is_configured": true, 00:08:42.311 "data_offset": 2048, 00:08:42.311 "data_size": 63488 00:08:42.311 }, 00:08:42.311 { 00:08:42.311 "name": "pt3", 00:08:42.311 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:42.311 "is_configured": true, 00:08:42.311 "data_offset": 2048, 00:08:42.311 "data_size": 63488 00:08:42.311 } 00:08:42.311 ] 00:08:42.311 }' 00:08:42.311 14:07:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.311 14:07:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.569 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.569 [2024-09-30 14:07:47.218372] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:42.829 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.829 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:42.829 "name": "raid_bdev1", 00:08:42.829 "aliases": [ 00:08:42.829 "6a3abf78-bbb1-4511-8b96-1d6840448253" 00:08:42.829 ], 00:08:42.829 "product_name": "Raid Volume", 00:08:42.829 "block_size": 512, 00:08:42.829 "num_blocks": 190464, 00:08:42.829 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:42.829 "assigned_rate_limits": { 00:08:42.829 "rw_ios_per_sec": 0, 00:08:42.829 "rw_mbytes_per_sec": 0, 00:08:42.829 "r_mbytes_per_sec": 0, 00:08:42.829 "w_mbytes_per_sec": 0 00:08:42.829 }, 00:08:42.829 "claimed": false, 00:08:42.829 "zoned": false, 00:08:42.829 "supported_io_types": { 00:08:42.829 "read": true, 00:08:42.829 "write": true, 00:08:42.829 "unmap": true, 00:08:42.829 "flush": true, 00:08:42.829 "reset": true, 00:08:42.829 "nvme_admin": false, 00:08:42.829 "nvme_io": false, 00:08:42.829 "nvme_io_md": false, 00:08:42.829 "write_zeroes": true, 00:08:42.829 "zcopy": false, 00:08:42.829 "get_zone_info": false, 00:08:42.829 "zone_management": false, 00:08:42.829 "zone_append": false, 00:08:42.829 "compare": false, 00:08:42.829 "compare_and_write": false, 00:08:42.829 "abort": false, 00:08:42.829 "seek_hole": false, 00:08:42.829 "seek_data": false, 00:08:42.829 "copy": false, 00:08:42.829 "nvme_iov_md": false 00:08:42.829 }, 00:08:42.829 "memory_domains": [ 00:08:42.829 { 00:08:42.829 "dma_device_id": "system", 00:08:42.829 "dma_device_type": 1 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.829 "dma_device_type": 2 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "dma_device_id": "system", 00:08:42.829 "dma_device_type": 1 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.829 "dma_device_type": 2 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "dma_device_id": "system", 00:08:42.829 "dma_device_type": 1 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.829 "dma_device_type": 2 00:08:42.829 } 00:08:42.829 ], 00:08:42.829 "driver_specific": { 00:08:42.829 "raid": { 00:08:42.829 "uuid": "6a3abf78-bbb1-4511-8b96-1d6840448253", 00:08:42.829 "strip_size_kb": 64, 00:08:42.829 "state": "online", 00:08:42.829 "raid_level": "concat", 00:08:42.829 "superblock": true, 00:08:42.829 "num_base_bdevs": 3, 00:08:42.829 "num_base_bdevs_discovered": 3, 00:08:42.829 "num_base_bdevs_operational": 3, 00:08:42.829 "base_bdevs_list": [ 00:08:42.829 { 00:08:42.829 "name": "pt1", 00:08:42.829 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:42.829 "is_configured": true, 00:08:42.829 "data_offset": 2048, 00:08:42.829 "data_size": 63488 00:08:42.829 }, 00:08:42.829 { 00:08:42.829 "name": "pt2", 00:08:42.830 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:42.830 "is_configured": true, 00:08:42.830 "data_offset": 2048, 00:08:42.830 "data_size": 63488 00:08:42.830 }, 00:08:42.830 { 00:08:42.830 "name": "pt3", 00:08:42.830 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:42.830 "is_configured": true, 00:08:42.830 "data_offset": 2048, 00:08:42.830 "data_size": 63488 00:08:42.830 } 00:08:42.830 ] 00:08:42.830 } 00:08:42.830 } 00:08:42.830 }' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:42.830 pt2 00:08:42.830 pt3' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.830 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:43.089 [2024-09-30 14:07:47.493857] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 6a3abf78-bbb1-4511-8b96-1d6840448253 '!=' 6a3abf78-bbb1-4511-8b96-1d6840448253 ']' 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 78986 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 78986 ']' 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 78986 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78986 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78986' 00:08:43.089 killing process with pid 78986 00:08:43.089 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 78986 00:08:43.089 [2024-09-30 14:07:47.573955] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:43.090 [2024-09-30 14:07:47.574091] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:43.090 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 78986 00:08:43.090 [2024-09-30 14:07:47.574173] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:43.090 [2024-09-30 14:07:47.574189] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:08:43.090 [2024-09-30 14:07:47.606185] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:43.349 14:07:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:43.349 ************************************ 00:08:43.349 END TEST raid_superblock_test 00:08:43.349 ************************************ 00:08:43.349 00:08:43.349 real 0m4.052s 00:08:43.349 user 0m6.428s 00:08:43.349 sys 0m0.833s 00:08:43.349 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:43.349 14:07:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.349 14:07:47 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:08:43.349 14:07:47 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:43.349 14:07:47 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:43.349 14:07:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:43.349 ************************************ 00:08:43.349 START TEST raid_read_error_test 00:08:43.349 ************************************ 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 read 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.KdrrCLxTqa 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79228 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79228 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 79228 ']' 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:43.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:43.349 14:07:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.607 [2024-09-30 14:07:48.023559] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:43.607 [2024-09-30 14:07:48.023682] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79228 ] 00:08:43.608 [2024-09-30 14:07:48.154136] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:43.608 [2024-09-30 14:07:48.183620] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:43.608 [2024-09-30 14:07:48.228709] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.867 [2024-09-30 14:07:48.270689] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.867 [2024-09-30 14:07:48.270724] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.434 BaseBdev1_malloc 00:08:44.434 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 true 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 [2024-09-30 14:07:48.876418] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:44.435 [2024-09-30 14:07:48.876570] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.435 [2024-09-30 14:07:48.876602] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:44.435 [2024-09-30 14:07:48.876635] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.435 [2024-09-30 14:07:48.878685] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.435 [2024-09-30 14:07:48.878757] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:44.435 BaseBdev1 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 BaseBdev2_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 true 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 [2024-09-30 14:07:48.933391] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:44.435 [2024-09-30 14:07:48.933466] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.435 [2024-09-30 14:07:48.933509] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:44.435 [2024-09-30 14:07:48.933528] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.435 [2024-09-30 14:07:48.936750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.435 [2024-09-30 14:07:48.936810] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:44.435 BaseBdev2 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 BaseBdev3_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 true 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 [2024-09-30 14:07:48.973938] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:44.435 [2024-09-30 14:07:48.973985] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.435 [2024-09-30 14:07:48.973999] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:44.435 [2024-09-30 14:07:48.974009] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.435 [2024-09-30 14:07:48.975981] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.435 [2024-09-30 14:07:48.976019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:44.435 BaseBdev3 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 [2024-09-30 14:07:48.985985] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.435 [2024-09-30 14:07:48.987786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:44.435 [2024-09-30 14:07:48.987866] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:44.435 [2024-09-30 14:07:48.988039] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:44.435 [2024-09-30 14:07:48.988067] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:44.435 [2024-09-30 14:07:48.988286] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:44.435 [2024-09-30 14:07:48.988425] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:44.435 [2024-09-30 14:07:48.988450] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:08:44.435 [2024-09-30 14:07:48.988595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.435 14:07:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.435 14:07:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.435 14:07:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.435 "name": "raid_bdev1", 00:08:44.435 "uuid": "1cd531a4-6653-42cc-be90-7e5db9180c44", 00:08:44.435 "strip_size_kb": 64, 00:08:44.435 "state": "online", 00:08:44.435 "raid_level": "concat", 00:08:44.435 "superblock": true, 00:08:44.435 "num_base_bdevs": 3, 00:08:44.435 "num_base_bdevs_discovered": 3, 00:08:44.435 "num_base_bdevs_operational": 3, 00:08:44.435 "base_bdevs_list": [ 00:08:44.435 { 00:08:44.435 "name": "BaseBdev1", 00:08:44.435 "uuid": "948690b8-0acd-54a7-a65c-827bfc1aa4e6", 00:08:44.435 "is_configured": true, 00:08:44.435 "data_offset": 2048, 00:08:44.435 "data_size": 63488 00:08:44.435 }, 00:08:44.435 { 00:08:44.435 "name": "BaseBdev2", 00:08:44.435 "uuid": "c6463cc4-e500-5c8d-a31a-5ae88da582d7", 00:08:44.435 "is_configured": true, 00:08:44.435 "data_offset": 2048, 00:08:44.435 "data_size": 63488 00:08:44.435 }, 00:08:44.435 { 00:08:44.435 "name": "BaseBdev3", 00:08:44.435 "uuid": "f912cc2d-9a62-5818-bd82-e23587064916", 00:08:44.435 "is_configured": true, 00:08:44.435 "data_offset": 2048, 00:08:44.435 "data_size": 63488 00:08:44.435 } 00:08:44.435 ] 00:08:44.435 }' 00:08:44.435 14:07:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.435 14:07:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.004 14:07:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:45.004 14:07:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:45.004 [2024-09-30 14:07:49.509467] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.943 "name": "raid_bdev1", 00:08:45.943 "uuid": "1cd531a4-6653-42cc-be90-7e5db9180c44", 00:08:45.943 "strip_size_kb": 64, 00:08:45.943 "state": "online", 00:08:45.943 "raid_level": "concat", 00:08:45.943 "superblock": true, 00:08:45.943 "num_base_bdevs": 3, 00:08:45.943 "num_base_bdevs_discovered": 3, 00:08:45.943 "num_base_bdevs_operational": 3, 00:08:45.943 "base_bdevs_list": [ 00:08:45.943 { 00:08:45.943 "name": "BaseBdev1", 00:08:45.943 "uuid": "948690b8-0acd-54a7-a65c-827bfc1aa4e6", 00:08:45.943 "is_configured": true, 00:08:45.943 "data_offset": 2048, 00:08:45.943 "data_size": 63488 00:08:45.943 }, 00:08:45.943 { 00:08:45.943 "name": "BaseBdev2", 00:08:45.943 "uuid": "c6463cc4-e500-5c8d-a31a-5ae88da582d7", 00:08:45.943 "is_configured": true, 00:08:45.943 "data_offset": 2048, 00:08:45.943 "data_size": 63488 00:08:45.943 }, 00:08:45.943 { 00:08:45.943 "name": "BaseBdev3", 00:08:45.943 "uuid": "f912cc2d-9a62-5818-bd82-e23587064916", 00:08:45.943 "is_configured": true, 00:08:45.943 "data_offset": 2048, 00:08:45.943 "data_size": 63488 00:08:45.943 } 00:08:45.943 ] 00:08:45.943 }' 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.943 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.511 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:46.511 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.511 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.511 [2024-09-30 14:07:50.873198] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:46.511 [2024-09-30 14:07:50.873246] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:46.511 [2024-09-30 14:07:50.875683] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:46.511 [2024-09-30 14:07:50.875733] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.511 [2024-09-30 14:07:50.875770] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:46.511 [2024-09-30 14:07:50.875780] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:08:46.511 { 00:08:46.511 "results": [ 00:08:46.511 { 00:08:46.511 "job": "raid_bdev1", 00:08:46.511 "core_mask": "0x1", 00:08:46.511 "workload": "randrw", 00:08:46.511 "percentage": 50, 00:08:46.511 "status": "finished", 00:08:46.511 "queue_depth": 1, 00:08:46.512 "io_size": 131072, 00:08:46.512 "runtime": 1.364587, 00:08:46.512 "iops": 17325.38856078799, 00:08:46.512 "mibps": 2165.6735700984987, 00:08:46.512 "io_failed": 1, 00:08:46.512 "io_timeout": 0, 00:08:46.512 "avg_latency_us": 80.03782199075883, 00:08:46.512 "min_latency_us": 24.593886462882097, 00:08:46.512 "max_latency_us": 1380.8349344978167 00:08:46.512 } 00:08:46.512 ], 00:08:46.512 "core_count": 1 00:08:46.512 } 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79228 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 79228 ']' 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 79228 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79228 00:08:46.512 killing process with pid 79228 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79228' 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 79228 00:08:46.512 [2024-09-30 14:07:50.923138] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:46.512 14:07:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 79228 00:08:46.512 [2024-09-30 14:07:50.948443] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.KdrrCLxTqa 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:46.771 00:08:46.771 real 0m3.273s 00:08:46.771 user 0m4.083s 00:08:46.771 sys 0m0.529s 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:46.771 14:07:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.771 ************************************ 00:08:46.771 END TEST raid_read_error_test 00:08:46.771 ************************************ 00:08:46.771 14:07:51 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:08:46.771 14:07:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:46.771 14:07:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:46.771 14:07:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:46.771 ************************************ 00:08:46.771 START TEST raid_write_error_test 00:08:46.771 ************************************ 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 write 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.w2KKXuhOSq 00:08:46.771 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79357 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79357 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 79357 ']' 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:46.772 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:46.772 14:07:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.772 [2024-09-30 14:07:51.360475] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:46.772 [2024-09-30 14:07:51.360607] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79357 ] 00:08:47.029 [2024-09-30 14:07:51.491776] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:47.029 [2024-09-30 14:07:51.516611] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:47.029 [2024-09-30 14:07:51.560436] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.029 [2024-09-30 14:07:51.601238] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.029 [2024-09-30 14:07:51.601275] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.597 BaseBdev1_malloc 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.597 true 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.597 [2024-09-30 14:07:52.246562] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:47.597 [2024-09-30 14:07:52.246624] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.597 [2024-09-30 14:07:52.246646] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:47.597 [2024-09-30 14:07:52.246674] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.597 [2024-09-30 14:07:52.248825] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.597 [2024-09-30 14:07:52.248865] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:47.597 BaseBdev1 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:47.597 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 BaseBdev2_malloc 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 true 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 [2024-09-30 14:07:52.297060] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:47.856 [2024-09-30 14:07:52.297122] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.856 [2024-09-30 14:07:52.297140] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:47.856 [2024-09-30 14:07:52.297149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.856 [2024-09-30 14:07:52.299102] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.856 [2024-09-30 14:07:52.299144] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:47.856 BaseBdev2 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 BaseBdev3_malloc 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 true 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 [2024-09-30 14:07:52.337318] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:47.856 [2024-09-30 14:07:52.337373] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.856 [2024-09-30 14:07:52.337388] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:47.856 [2024-09-30 14:07:52.337398] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.856 [2024-09-30 14:07:52.339350] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.856 [2024-09-30 14:07:52.339388] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:47.856 BaseBdev3 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.856 [2024-09-30 14:07:52.349377] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:47.856 [2024-09-30 14:07:52.351116] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:47.856 [2024-09-30 14:07:52.351191] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:47.856 [2024-09-30 14:07:52.351352] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:47.856 [2024-09-30 14:07:52.351369] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:47.856 [2024-09-30 14:07:52.351603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:47.856 [2024-09-30 14:07:52.351729] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:47.856 [2024-09-30 14:07:52.351747] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:08:47.856 [2024-09-30 14:07:52.351858] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.856 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.857 "name": "raid_bdev1", 00:08:47.857 "uuid": "12e232c8-e2be-4d71-8bbb-e805902dc6dd", 00:08:47.857 "strip_size_kb": 64, 00:08:47.857 "state": "online", 00:08:47.857 "raid_level": "concat", 00:08:47.857 "superblock": true, 00:08:47.857 "num_base_bdevs": 3, 00:08:47.857 "num_base_bdevs_discovered": 3, 00:08:47.857 "num_base_bdevs_operational": 3, 00:08:47.857 "base_bdevs_list": [ 00:08:47.857 { 00:08:47.857 "name": "BaseBdev1", 00:08:47.857 "uuid": "c8274169-6e0b-5f29-af44-d8de41f1240b", 00:08:47.857 "is_configured": true, 00:08:47.857 "data_offset": 2048, 00:08:47.857 "data_size": 63488 00:08:47.857 }, 00:08:47.857 { 00:08:47.857 "name": "BaseBdev2", 00:08:47.857 "uuid": "6ae08471-1154-50b7-afe0-8424a95d8833", 00:08:47.857 "is_configured": true, 00:08:47.857 "data_offset": 2048, 00:08:47.857 "data_size": 63488 00:08:47.857 }, 00:08:47.857 { 00:08:47.857 "name": "BaseBdev3", 00:08:47.857 "uuid": "fbb796e5-d53b-504e-bdaf-9444b2ffb754", 00:08:47.857 "is_configured": true, 00:08:47.857 "data_offset": 2048, 00:08:47.857 "data_size": 63488 00:08:47.857 } 00:08:47.857 ] 00:08:47.857 }' 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.857 14:07:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.425 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:48.425 14:07:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:48.425 [2024-09-30 14:07:52.868941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.363 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.364 "name": "raid_bdev1", 00:08:49.364 "uuid": "12e232c8-e2be-4d71-8bbb-e805902dc6dd", 00:08:49.364 "strip_size_kb": 64, 00:08:49.364 "state": "online", 00:08:49.364 "raid_level": "concat", 00:08:49.364 "superblock": true, 00:08:49.364 "num_base_bdevs": 3, 00:08:49.364 "num_base_bdevs_discovered": 3, 00:08:49.364 "num_base_bdevs_operational": 3, 00:08:49.364 "base_bdevs_list": [ 00:08:49.364 { 00:08:49.364 "name": "BaseBdev1", 00:08:49.364 "uuid": "c8274169-6e0b-5f29-af44-d8de41f1240b", 00:08:49.364 "is_configured": true, 00:08:49.364 "data_offset": 2048, 00:08:49.364 "data_size": 63488 00:08:49.364 }, 00:08:49.364 { 00:08:49.364 "name": "BaseBdev2", 00:08:49.364 "uuid": "6ae08471-1154-50b7-afe0-8424a95d8833", 00:08:49.364 "is_configured": true, 00:08:49.364 "data_offset": 2048, 00:08:49.364 "data_size": 63488 00:08:49.364 }, 00:08:49.364 { 00:08:49.364 "name": "BaseBdev3", 00:08:49.364 "uuid": "fbb796e5-d53b-504e-bdaf-9444b2ffb754", 00:08:49.364 "is_configured": true, 00:08:49.364 "data_offset": 2048, 00:08:49.364 "data_size": 63488 00:08:49.364 } 00:08:49.364 ] 00:08:49.364 }' 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.364 14:07:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.623 [2024-09-30 14:07:54.236468] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:49.623 [2024-09-30 14:07:54.236608] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:49.623 [2024-09-30 14:07:54.239039] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:49.623 [2024-09-30 14:07:54.239125] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.623 [2024-09-30 14:07:54.239166] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:49.623 [2024-09-30 14:07:54.239174] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:08:49.623 { 00:08:49.623 "results": [ 00:08:49.623 { 00:08:49.623 "job": "raid_bdev1", 00:08:49.623 "core_mask": "0x1", 00:08:49.623 "workload": "randrw", 00:08:49.623 "percentage": 50, 00:08:49.623 "status": "finished", 00:08:49.623 "queue_depth": 1, 00:08:49.623 "io_size": 131072, 00:08:49.623 "runtime": 1.368474, 00:08:49.623 "iops": 17349.982535291136, 00:08:49.623 "mibps": 2168.747816911392, 00:08:49.623 "io_failed": 1, 00:08:49.623 "io_timeout": 0, 00:08:49.623 "avg_latency_us": 79.8478419002107, 00:08:49.623 "min_latency_us": 24.258515283842794, 00:08:49.623 "max_latency_us": 1323.598253275109 00:08:49.623 } 00:08:49.623 ], 00:08:49.623 "core_count": 1 00:08:49.623 } 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79357 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 79357 ']' 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 79357 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:49.623 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79357 00:08:49.883 killing process with pid 79357 00:08:49.883 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:49.883 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:49.883 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79357' 00:08:49.883 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 79357 00:08:49.883 [2024-09-30 14:07:54.284295] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:49.883 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 79357 00:08:49.883 [2024-09-30 14:07:54.308815] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.w2KKXuhOSq 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:50.143 00:08:50.143 real 0m3.289s 00:08:50.143 user 0m4.155s 00:08:50.143 sys 0m0.519s 00:08:50.143 ************************************ 00:08:50.143 END TEST raid_write_error_test 00:08:50.143 ************************************ 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:50.143 14:07:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.143 14:07:54 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:50.143 14:07:54 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:08:50.143 14:07:54 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:50.143 14:07:54 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:50.143 14:07:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:50.143 ************************************ 00:08:50.143 START TEST raid_state_function_test 00:08:50.143 ************************************ 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 false 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:50.143 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=79484 00:08:50.144 Process raid pid: 79484 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 79484' 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 79484 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 79484 ']' 00:08:50.144 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:50.144 14:07:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.144 [2024-09-30 14:07:54.717775] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:50.144 [2024-09-30 14:07:54.717892] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:50.403 [2024-09-30 14:07:54.849469] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:50.403 [2024-09-30 14:07:54.879183] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:50.403 [2024-09-30 14:07:54.923866] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:50.403 [2024-09-30 14:07:54.964887] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:50.403 [2024-09-30 14:07:54.964921] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.971 [2024-09-30 14:07:55.553652] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:50.971 [2024-09-30 14:07:55.553774] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:50.971 [2024-09-30 14:07:55.553822] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:50.971 [2024-09-30 14:07:55.553844] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:50.971 [2024-09-30 14:07:55.553866] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:50.971 [2024-09-30 14:07:55.553883] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.971 "name": "Existed_Raid", 00:08:50.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.971 "strip_size_kb": 0, 00:08:50.971 "state": "configuring", 00:08:50.971 "raid_level": "raid1", 00:08:50.971 "superblock": false, 00:08:50.971 "num_base_bdevs": 3, 00:08:50.971 "num_base_bdevs_discovered": 0, 00:08:50.971 "num_base_bdevs_operational": 3, 00:08:50.971 "base_bdevs_list": [ 00:08:50.971 { 00:08:50.971 "name": "BaseBdev1", 00:08:50.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.971 "is_configured": false, 00:08:50.971 "data_offset": 0, 00:08:50.971 "data_size": 0 00:08:50.971 }, 00:08:50.971 { 00:08:50.971 "name": "BaseBdev2", 00:08:50.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.971 "is_configured": false, 00:08:50.971 "data_offset": 0, 00:08:50.971 "data_size": 0 00:08:50.971 }, 00:08:50.971 { 00:08:50.971 "name": "BaseBdev3", 00:08:50.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.971 "is_configured": false, 00:08:50.971 "data_offset": 0, 00:08:50.971 "data_size": 0 00:08:50.971 } 00:08:50.971 ] 00:08:50.971 }' 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.971 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.538 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:51.538 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 [2024-09-30 14:07:55.948922] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:51.539 [2024-09-30 14:07:55.949008] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 [2024-09-30 14:07:55.960906] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:51.539 [2024-09-30 14:07:55.960983] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:51.539 [2024-09-30 14:07:55.961027] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:51.539 [2024-09-30 14:07:55.961048] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:51.539 [2024-09-30 14:07:55.961066] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:51.539 [2024-09-30 14:07:55.961085] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 [2024-09-30 14:07:55.981565] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:51.539 BaseBdev1 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 [ 00:08:51.539 { 00:08:51.539 "name": "BaseBdev1", 00:08:51.539 "aliases": [ 00:08:51.539 "7acb8ffd-8242-4f13-a633-587f3d42b11a" 00:08:51.539 ], 00:08:51.539 "product_name": "Malloc disk", 00:08:51.539 "block_size": 512, 00:08:51.539 "num_blocks": 65536, 00:08:51.539 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:51.539 "assigned_rate_limits": { 00:08:51.539 "rw_ios_per_sec": 0, 00:08:51.539 "rw_mbytes_per_sec": 0, 00:08:51.539 "r_mbytes_per_sec": 0, 00:08:51.539 "w_mbytes_per_sec": 0 00:08:51.539 }, 00:08:51.539 "claimed": true, 00:08:51.539 "claim_type": "exclusive_write", 00:08:51.539 "zoned": false, 00:08:51.539 "supported_io_types": { 00:08:51.539 "read": true, 00:08:51.539 "write": true, 00:08:51.539 "unmap": true, 00:08:51.539 "flush": true, 00:08:51.539 "reset": true, 00:08:51.539 "nvme_admin": false, 00:08:51.539 "nvme_io": false, 00:08:51.539 "nvme_io_md": false, 00:08:51.539 "write_zeroes": true, 00:08:51.539 "zcopy": true, 00:08:51.539 "get_zone_info": false, 00:08:51.539 "zone_management": false, 00:08:51.539 "zone_append": false, 00:08:51.539 "compare": false, 00:08:51.539 "compare_and_write": false, 00:08:51.539 "abort": true, 00:08:51.539 "seek_hole": false, 00:08:51.539 "seek_data": false, 00:08:51.539 "copy": true, 00:08:51.539 "nvme_iov_md": false 00:08:51.539 }, 00:08:51.539 "memory_domains": [ 00:08:51.539 { 00:08:51.539 "dma_device_id": "system", 00:08:51.539 "dma_device_type": 1 00:08:51.539 }, 00:08:51.539 { 00:08:51.539 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.539 "dma_device_type": 2 00:08:51.539 } 00:08:51.539 ], 00:08:51.539 "driver_specific": {} 00:08:51.539 } 00:08:51.539 ] 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.539 "name": "Existed_Raid", 00:08:51.539 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.539 "strip_size_kb": 0, 00:08:51.539 "state": "configuring", 00:08:51.539 "raid_level": "raid1", 00:08:51.539 "superblock": false, 00:08:51.539 "num_base_bdevs": 3, 00:08:51.539 "num_base_bdevs_discovered": 1, 00:08:51.539 "num_base_bdevs_operational": 3, 00:08:51.539 "base_bdevs_list": [ 00:08:51.539 { 00:08:51.539 "name": "BaseBdev1", 00:08:51.539 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:51.539 "is_configured": true, 00:08:51.539 "data_offset": 0, 00:08:51.539 "data_size": 65536 00:08:51.539 }, 00:08:51.539 { 00:08:51.539 "name": "BaseBdev2", 00:08:51.539 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.539 "is_configured": false, 00:08:51.539 "data_offset": 0, 00:08:51.539 "data_size": 0 00:08:51.539 }, 00:08:51.539 { 00:08:51.539 "name": "BaseBdev3", 00:08:51.539 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.539 "is_configured": false, 00:08:51.539 "data_offset": 0, 00:08:51.539 "data_size": 0 00:08:51.539 } 00:08:51.539 ] 00:08:51.539 }' 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.539 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.108 [2024-09-30 14:07:56.476707] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:52.108 [2024-09-30 14:07:56.476810] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.108 [2024-09-30 14:07:56.488713] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.108 [2024-09-30 14:07:56.490528] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:52.108 [2024-09-30 14:07:56.490597] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:52.108 [2024-09-30 14:07:56.490627] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:52.108 [2024-09-30 14:07:56.490653] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.108 "name": "Existed_Raid", 00:08:52.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.108 "strip_size_kb": 0, 00:08:52.108 "state": "configuring", 00:08:52.108 "raid_level": "raid1", 00:08:52.108 "superblock": false, 00:08:52.108 "num_base_bdevs": 3, 00:08:52.108 "num_base_bdevs_discovered": 1, 00:08:52.108 "num_base_bdevs_operational": 3, 00:08:52.108 "base_bdevs_list": [ 00:08:52.108 { 00:08:52.108 "name": "BaseBdev1", 00:08:52.108 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:52.108 "is_configured": true, 00:08:52.108 "data_offset": 0, 00:08:52.108 "data_size": 65536 00:08:52.108 }, 00:08:52.108 { 00:08:52.108 "name": "BaseBdev2", 00:08:52.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.108 "is_configured": false, 00:08:52.108 "data_offset": 0, 00:08:52.108 "data_size": 0 00:08:52.108 }, 00:08:52.108 { 00:08:52.108 "name": "BaseBdev3", 00:08:52.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.108 "is_configured": false, 00:08:52.108 "data_offset": 0, 00:08:52.108 "data_size": 0 00:08:52.108 } 00:08:52.108 ] 00:08:52.108 }' 00:08:52.108 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.109 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.368 [2024-09-30 14:07:56.954710] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:52.368 BaseBdev2 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.368 [ 00:08:52.368 { 00:08:52.368 "name": "BaseBdev2", 00:08:52.368 "aliases": [ 00:08:52.368 "2db92b71-1aba-4d3e-ac26-37140adcaff3" 00:08:52.368 ], 00:08:52.368 "product_name": "Malloc disk", 00:08:52.368 "block_size": 512, 00:08:52.368 "num_blocks": 65536, 00:08:52.368 "uuid": "2db92b71-1aba-4d3e-ac26-37140adcaff3", 00:08:52.368 "assigned_rate_limits": { 00:08:52.368 "rw_ios_per_sec": 0, 00:08:52.368 "rw_mbytes_per_sec": 0, 00:08:52.368 "r_mbytes_per_sec": 0, 00:08:52.368 "w_mbytes_per_sec": 0 00:08:52.368 }, 00:08:52.368 "claimed": true, 00:08:52.368 "claim_type": "exclusive_write", 00:08:52.368 "zoned": false, 00:08:52.368 "supported_io_types": { 00:08:52.368 "read": true, 00:08:52.368 "write": true, 00:08:52.368 "unmap": true, 00:08:52.368 "flush": true, 00:08:52.368 "reset": true, 00:08:52.368 "nvme_admin": false, 00:08:52.368 "nvme_io": false, 00:08:52.368 "nvme_io_md": false, 00:08:52.368 "write_zeroes": true, 00:08:52.368 "zcopy": true, 00:08:52.368 "get_zone_info": false, 00:08:52.368 "zone_management": false, 00:08:52.368 "zone_append": false, 00:08:52.368 "compare": false, 00:08:52.368 "compare_and_write": false, 00:08:52.368 "abort": true, 00:08:52.368 "seek_hole": false, 00:08:52.368 "seek_data": false, 00:08:52.368 "copy": true, 00:08:52.368 "nvme_iov_md": false 00:08:52.368 }, 00:08:52.368 "memory_domains": [ 00:08:52.368 { 00:08:52.368 "dma_device_id": "system", 00:08:52.368 "dma_device_type": 1 00:08:52.368 }, 00:08:52.368 { 00:08:52.368 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.368 "dma_device_type": 2 00:08:52.368 } 00:08:52.368 ], 00:08:52.368 "driver_specific": {} 00:08:52.368 } 00:08:52.368 ] 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.368 14:07:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.368 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.628 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.628 "name": "Existed_Raid", 00:08:52.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.628 "strip_size_kb": 0, 00:08:52.628 "state": "configuring", 00:08:52.628 "raid_level": "raid1", 00:08:52.628 "superblock": false, 00:08:52.628 "num_base_bdevs": 3, 00:08:52.628 "num_base_bdevs_discovered": 2, 00:08:52.628 "num_base_bdevs_operational": 3, 00:08:52.628 "base_bdevs_list": [ 00:08:52.628 { 00:08:52.628 "name": "BaseBdev1", 00:08:52.628 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:52.628 "is_configured": true, 00:08:52.628 "data_offset": 0, 00:08:52.628 "data_size": 65536 00:08:52.628 }, 00:08:52.628 { 00:08:52.628 "name": "BaseBdev2", 00:08:52.628 "uuid": "2db92b71-1aba-4d3e-ac26-37140adcaff3", 00:08:52.628 "is_configured": true, 00:08:52.628 "data_offset": 0, 00:08:52.628 "data_size": 65536 00:08:52.628 }, 00:08:52.628 { 00:08:52.628 "name": "BaseBdev3", 00:08:52.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.628 "is_configured": false, 00:08:52.628 "data_offset": 0, 00:08:52.628 "data_size": 0 00:08:52.628 } 00:08:52.628 ] 00:08:52.628 }' 00:08:52.628 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.628 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.888 [2024-09-30 14:07:57.432729] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:52.888 [2024-09-30 14:07:57.432847] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:52.888 [2024-09-30 14:07:57.432870] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:52.888 [2024-09-30 14:07:57.433137] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:52.888 [2024-09-30 14:07:57.433315] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:52.888 [2024-09-30 14:07:57.433357] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:52.888 [2024-09-30 14:07:57.433590] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:52.888 BaseBdev3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.888 [ 00:08:52.888 { 00:08:52.888 "name": "BaseBdev3", 00:08:52.888 "aliases": [ 00:08:52.888 "bffe0a63-477a-4586-b348-3013ba57165c" 00:08:52.888 ], 00:08:52.888 "product_name": "Malloc disk", 00:08:52.888 "block_size": 512, 00:08:52.888 "num_blocks": 65536, 00:08:52.888 "uuid": "bffe0a63-477a-4586-b348-3013ba57165c", 00:08:52.888 "assigned_rate_limits": { 00:08:52.888 "rw_ios_per_sec": 0, 00:08:52.888 "rw_mbytes_per_sec": 0, 00:08:52.888 "r_mbytes_per_sec": 0, 00:08:52.888 "w_mbytes_per_sec": 0 00:08:52.888 }, 00:08:52.888 "claimed": true, 00:08:52.888 "claim_type": "exclusive_write", 00:08:52.888 "zoned": false, 00:08:52.888 "supported_io_types": { 00:08:52.888 "read": true, 00:08:52.888 "write": true, 00:08:52.888 "unmap": true, 00:08:52.888 "flush": true, 00:08:52.888 "reset": true, 00:08:52.888 "nvme_admin": false, 00:08:52.888 "nvme_io": false, 00:08:52.888 "nvme_io_md": false, 00:08:52.888 "write_zeroes": true, 00:08:52.888 "zcopy": true, 00:08:52.888 "get_zone_info": false, 00:08:52.888 "zone_management": false, 00:08:52.888 "zone_append": false, 00:08:52.888 "compare": false, 00:08:52.888 "compare_and_write": false, 00:08:52.888 "abort": true, 00:08:52.888 "seek_hole": false, 00:08:52.888 "seek_data": false, 00:08:52.888 "copy": true, 00:08:52.888 "nvme_iov_md": false 00:08:52.888 }, 00:08:52.888 "memory_domains": [ 00:08:52.888 { 00:08:52.888 "dma_device_id": "system", 00:08:52.888 "dma_device_type": 1 00:08:52.888 }, 00:08:52.888 { 00:08:52.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.888 "dma_device_type": 2 00:08:52.888 } 00:08:52.888 ], 00:08:52.888 "driver_specific": {} 00:08:52.888 } 00:08:52.888 ] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.888 "name": "Existed_Raid", 00:08:52.888 "uuid": "7408bcc2-821c-409f-af72-af3c53370ebf", 00:08:52.888 "strip_size_kb": 0, 00:08:52.888 "state": "online", 00:08:52.888 "raid_level": "raid1", 00:08:52.888 "superblock": false, 00:08:52.888 "num_base_bdevs": 3, 00:08:52.888 "num_base_bdevs_discovered": 3, 00:08:52.888 "num_base_bdevs_operational": 3, 00:08:52.888 "base_bdevs_list": [ 00:08:52.888 { 00:08:52.888 "name": "BaseBdev1", 00:08:52.888 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:52.888 "is_configured": true, 00:08:52.888 "data_offset": 0, 00:08:52.888 "data_size": 65536 00:08:52.888 }, 00:08:52.888 { 00:08:52.888 "name": "BaseBdev2", 00:08:52.888 "uuid": "2db92b71-1aba-4d3e-ac26-37140adcaff3", 00:08:52.888 "is_configured": true, 00:08:52.888 "data_offset": 0, 00:08:52.888 "data_size": 65536 00:08:52.888 }, 00:08:52.888 { 00:08:52.888 "name": "BaseBdev3", 00:08:52.888 "uuid": "bffe0a63-477a-4586-b348-3013ba57165c", 00:08:52.888 "is_configured": true, 00:08:52.888 "data_offset": 0, 00:08:52.888 "data_size": 65536 00:08:52.888 } 00:08:52.888 ] 00:08:52.888 }' 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.888 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.456 [2024-09-30 14:07:57.920225] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.456 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:53.456 "name": "Existed_Raid", 00:08:53.456 "aliases": [ 00:08:53.456 "7408bcc2-821c-409f-af72-af3c53370ebf" 00:08:53.456 ], 00:08:53.456 "product_name": "Raid Volume", 00:08:53.456 "block_size": 512, 00:08:53.456 "num_blocks": 65536, 00:08:53.456 "uuid": "7408bcc2-821c-409f-af72-af3c53370ebf", 00:08:53.456 "assigned_rate_limits": { 00:08:53.456 "rw_ios_per_sec": 0, 00:08:53.456 "rw_mbytes_per_sec": 0, 00:08:53.456 "r_mbytes_per_sec": 0, 00:08:53.456 "w_mbytes_per_sec": 0 00:08:53.456 }, 00:08:53.456 "claimed": false, 00:08:53.456 "zoned": false, 00:08:53.456 "supported_io_types": { 00:08:53.456 "read": true, 00:08:53.456 "write": true, 00:08:53.456 "unmap": false, 00:08:53.456 "flush": false, 00:08:53.456 "reset": true, 00:08:53.456 "nvme_admin": false, 00:08:53.456 "nvme_io": false, 00:08:53.456 "nvme_io_md": false, 00:08:53.456 "write_zeroes": true, 00:08:53.456 "zcopy": false, 00:08:53.456 "get_zone_info": false, 00:08:53.456 "zone_management": false, 00:08:53.456 "zone_append": false, 00:08:53.456 "compare": false, 00:08:53.456 "compare_and_write": false, 00:08:53.456 "abort": false, 00:08:53.456 "seek_hole": false, 00:08:53.456 "seek_data": false, 00:08:53.456 "copy": false, 00:08:53.456 "nvme_iov_md": false 00:08:53.456 }, 00:08:53.456 "memory_domains": [ 00:08:53.456 { 00:08:53.456 "dma_device_id": "system", 00:08:53.456 "dma_device_type": 1 00:08:53.456 }, 00:08:53.456 { 00:08:53.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.456 "dma_device_type": 2 00:08:53.456 }, 00:08:53.457 { 00:08:53.457 "dma_device_id": "system", 00:08:53.457 "dma_device_type": 1 00:08:53.457 }, 00:08:53.457 { 00:08:53.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.457 "dma_device_type": 2 00:08:53.457 }, 00:08:53.457 { 00:08:53.457 "dma_device_id": "system", 00:08:53.457 "dma_device_type": 1 00:08:53.457 }, 00:08:53.457 { 00:08:53.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.457 "dma_device_type": 2 00:08:53.457 } 00:08:53.457 ], 00:08:53.457 "driver_specific": { 00:08:53.457 "raid": { 00:08:53.457 "uuid": "7408bcc2-821c-409f-af72-af3c53370ebf", 00:08:53.457 "strip_size_kb": 0, 00:08:53.457 "state": "online", 00:08:53.457 "raid_level": "raid1", 00:08:53.457 "superblock": false, 00:08:53.457 "num_base_bdevs": 3, 00:08:53.457 "num_base_bdevs_discovered": 3, 00:08:53.457 "num_base_bdevs_operational": 3, 00:08:53.457 "base_bdevs_list": [ 00:08:53.457 { 00:08:53.457 "name": "BaseBdev1", 00:08:53.457 "uuid": "7acb8ffd-8242-4f13-a633-587f3d42b11a", 00:08:53.457 "is_configured": true, 00:08:53.457 "data_offset": 0, 00:08:53.457 "data_size": 65536 00:08:53.457 }, 00:08:53.457 { 00:08:53.457 "name": "BaseBdev2", 00:08:53.457 "uuid": "2db92b71-1aba-4d3e-ac26-37140adcaff3", 00:08:53.457 "is_configured": true, 00:08:53.457 "data_offset": 0, 00:08:53.457 "data_size": 65536 00:08:53.457 }, 00:08:53.457 { 00:08:53.457 "name": "BaseBdev3", 00:08:53.457 "uuid": "bffe0a63-477a-4586-b348-3013ba57165c", 00:08:53.457 "is_configured": true, 00:08:53.457 "data_offset": 0, 00:08:53.457 "data_size": 65536 00:08:53.457 } 00:08:53.457 ] 00:08:53.457 } 00:08:53.457 } 00:08:53.457 }' 00:08:53.457 14:07:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:53.457 BaseBdev2 00:08:53.457 BaseBdev3' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.457 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.716 [2024-09-30 14:07:58.163561] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:53.716 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.717 "name": "Existed_Raid", 00:08:53.717 "uuid": "7408bcc2-821c-409f-af72-af3c53370ebf", 00:08:53.717 "strip_size_kb": 0, 00:08:53.717 "state": "online", 00:08:53.717 "raid_level": "raid1", 00:08:53.717 "superblock": false, 00:08:53.717 "num_base_bdevs": 3, 00:08:53.717 "num_base_bdevs_discovered": 2, 00:08:53.717 "num_base_bdevs_operational": 2, 00:08:53.717 "base_bdevs_list": [ 00:08:53.717 { 00:08:53.717 "name": null, 00:08:53.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.717 "is_configured": false, 00:08:53.717 "data_offset": 0, 00:08:53.717 "data_size": 65536 00:08:53.717 }, 00:08:53.717 { 00:08:53.717 "name": "BaseBdev2", 00:08:53.717 "uuid": "2db92b71-1aba-4d3e-ac26-37140adcaff3", 00:08:53.717 "is_configured": true, 00:08:53.717 "data_offset": 0, 00:08:53.717 "data_size": 65536 00:08:53.717 }, 00:08:53.717 { 00:08:53.717 "name": "BaseBdev3", 00:08:53.717 "uuid": "bffe0a63-477a-4586-b348-3013ba57165c", 00:08:53.717 "is_configured": true, 00:08:53.717 "data_offset": 0, 00:08:53.717 "data_size": 65536 00:08:53.717 } 00:08:53.717 ] 00:08:53.717 }' 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.717 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.975 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:53.976 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:53.976 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.976 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:53.976 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.976 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 [2024-09-30 14:07:58.677899] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 [2024-09-30 14:07:58.728799] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:54.235 [2024-09-30 14:07:58.728935] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:54.235 [2024-09-30 14:07:58.740170] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:54.235 [2024-09-30 14:07:58.740284] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:54.235 [2024-09-30 14:07:58.740324] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.235 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.235 [ 00:08:54.235 { 00:08:54.235 "name": "BaseBdev2", 00:08:54.235 "aliases": [ 00:08:54.235 "8b673c82-0166-493a-80c8-81f8a58f1526" 00:08:54.235 ], 00:08:54.235 "product_name": "Malloc disk", 00:08:54.235 "block_size": 512, 00:08:54.235 "num_blocks": 65536, 00:08:54.235 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:54.235 "assigned_rate_limits": { 00:08:54.235 "rw_ios_per_sec": 0, 00:08:54.235 "rw_mbytes_per_sec": 0, 00:08:54.235 "r_mbytes_per_sec": 0, 00:08:54.235 "w_mbytes_per_sec": 0 00:08:54.235 }, 00:08:54.235 "claimed": false, 00:08:54.235 "zoned": false, 00:08:54.235 "supported_io_types": { 00:08:54.235 "read": true, 00:08:54.235 "write": true, 00:08:54.235 "unmap": true, 00:08:54.235 "flush": true, 00:08:54.235 "reset": true, 00:08:54.235 "nvme_admin": false, 00:08:54.235 "nvme_io": false, 00:08:54.235 "nvme_io_md": false, 00:08:54.235 "write_zeroes": true, 00:08:54.235 "zcopy": true, 00:08:54.235 "get_zone_info": false, 00:08:54.235 "zone_management": false, 00:08:54.235 "zone_append": false, 00:08:54.235 "compare": false, 00:08:54.236 "compare_and_write": false, 00:08:54.236 "abort": true, 00:08:54.236 "seek_hole": false, 00:08:54.236 "seek_data": false, 00:08:54.236 "copy": true, 00:08:54.236 "nvme_iov_md": false 00:08:54.236 }, 00:08:54.236 "memory_domains": [ 00:08:54.236 { 00:08:54.236 "dma_device_id": "system", 00:08:54.236 "dma_device_type": 1 00:08:54.236 }, 00:08:54.236 { 00:08:54.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.236 "dma_device_type": 2 00:08:54.236 } 00:08:54.236 ], 00:08:54.236 "driver_specific": {} 00:08:54.236 } 00:08:54.236 ] 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.236 BaseBdev3 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.236 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.236 [ 00:08:54.236 { 00:08:54.236 "name": "BaseBdev3", 00:08:54.236 "aliases": [ 00:08:54.236 "cac8d6bd-5496-4175-a4dd-f36c5f814cd2" 00:08:54.236 ], 00:08:54.236 "product_name": "Malloc disk", 00:08:54.236 "block_size": 512, 00:08:54.236 "num_blocks": 65536, 00:08:54.236 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:54.236 "assigned_rate_limits": { 00:08:54.236 "rw_ios_per_sec": 0, 00:08:54.236 "rw_mbytes_per_sec": 0, 00:08:54.236 "r_mbytes_per_sec": 0, 00:08:54.236 "w_mbytes_per_sec": 0 00:08:54.236 }, 00:08:54.236 "claimed": false, 00:08:54.236 "zoned": false, 00:08:54.236 "supported_io_types": { 00:08:54.236 "read": true, 00:08:54.236 "write": true, 00:08:54.236 "unmap": true, 00:08:54.236 "flush": true, 00:08:54.236 "reset": true, 00:08:54.236 "nvme_admin": false, 00:08:54.236 "nvme_io": false, 00:08:54.236 "nvme_io_md": false, 00:08:54.236 "write_zeroes": true, 00:08:54.236 "zcopy": true, 00:08:54.236 "get_zone_info": false, 00:08:54.498 "zone_management": false, 00:08:54.498 "zone_append": false, 00:08:54.498 "compare": false, 00:08:54.498 "compare_and_write": false, 00:08:54.498 "abort": true, 00:08:54.498 "seek_hole": false, 00:08:54.498 "seek_data": false, 00:08:54.498 "copy": true, 00:08:54.498 "nvme_iov_md": false 00:08:54.498 }, 00:08:54.498 "memory_domains": [ 00:08:54.498 { 00:08:54.498 "dma_device_id": "system", 00:08:54.498 "dma_device_type": 1 00:08:54.498 }, 00:08:54.498 { 00:08:54.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.498 "dma_device_type": 2 00:08:54.498 } 00:08:54.498 ], 00:08:54.498 "driver_specific": {} 00:08:54.498 } 00:08:54.498 ] 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.498 [2024-09-30 14:07:58.902742] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:54.498 [2024-09-30 14:07:58.902865] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:54.498 [2024-09-30 14:07:58.902921] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:54.498 [2024-09-30 14:07:58.904746] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.498 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.499 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.499 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.499 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.499 "name": "Existed_Raid", 00:08:54.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.499 "strip_size_kb": 0, 00:08:54.499 "state": "configuring", 00:08:54.499 "raid_level": "raid1", 00:08:54.499 "superblock": false, 00:08:54.499 "num_base_bdevs": 3, 00:08:54.499 "num_base_bdevs_discovered": 2, 00:08:54.499 "num_base_bdevs_operational": 3, 00:08:54.499 "base_bdevs_list": [ 00:08:54.499 { 00:08:54.499 "name": "BaseBdev1", 00:08:54.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.499 "is_configured": false, 00:08:54.499 "data_offset": 0, 00:08:54.499 "data_size": 0 00:08:54.499 }, 00:08:54.499 { 00:08:54.499 "name": "BaseBdev2", 00:08:54.499 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:54.499 "is_configured": true, 00:08:54.499 "data_offset": 0, 00:08:54.499 "data_size": 65536 00:08:54.499 }, 00:08:54.499 { 00:08:54.499 "name": "BaseBdev3", 00:08:54.499 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:54.499 "is_configured": true, 00:08:54.499 "data_offset": 0, 00:08:54.499 "data_size": 65536 00:08:54.499 } 00:08:54.499 ] 00:08:54.499 }' 00:08:54.499 14:07:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.499 14:07:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.763 [2024-09-30 14:07:59.325986] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.763 "name": "Existed_Raid", 00:08:54.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.763 "strip_size_kb": 0, 00:08:54.763 "state": "configuring", 00:08:54.763 "raid_level": "raid1", 00:08:54.763 "superblock": false, 00:08:54.763 "num_base_bdevs": 3, 00:08:54.763 "num_base_bdevs_discovered": 1, 00:08:54.763 "num_base_bdevs_operational": 3, 00:08:54.763 "base_bdevs_list": [ 00:08:54.763 { 00:08:54.763 "name": "BaseBdev1", 00:08:54.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.763 "is_configured": false, 00:08:54.763 "data_offset": 0, 00:08:54.763 "data_size": 0 00:08:54.763 }, 00:08:54.763 { 00:08:54.763 "name": null, 00:08:54.763 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:54.763 "is_configured": false, 00:08:54.763 "data_offset": 0, 00:08:54.763 "data_size": 65536 00:08:54.763 }, 00:08:54.763 { 00:08:54.763 "name": "BaseBdev3", 00:08:54.763 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:54.763 "is_configured": true, 00:08:54.763 "data_offset": 0, 00:08:54.763 "data_size": 65536 00:08:54.763 } 00:08:54.763 ] 00:08:54.763 }' 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.763 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.331 [2024-09-30 14:07:59.839824] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:55.331 BaseBdev1 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.331 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.331 [ 00:08:55.331 { 00:08:55.331 "name": "BaseBdev1", 00:08:55.331 "aliases": [ 00:08:55.331 "acf85c78-3dbd-4e33-acac-65e1e93fdf7b" 00:08:55.331 ], 00:08:55.331 "product_name": "Malloc disk", 00:08:55.331 "block_size": 512, 00:08:55.331 "num_blocks": 65536, 00:08:55.331 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:55.331 "assigned_rate_limits": { 00:08:55.331 "rw_ios_per_sec": 0, 00:08:55.331 "rw_mbytes_per_sec": 0, 00:08:55.331 "r_mbytes_per_sec": 0, 00:08:55.331 "w_mbytes_per_sec": 0 00:08:55.331 }, 00:08:55.331 "claimed": true, 00:08:55.331 "claim_type": "exclusive_write", 00:08:55.331 "zoned": false, 00:08:55.331 "supported_io_types": { 00:08:55.331 "read": true, 00:08:55.331 "write": true, 00:08:55.331 "unmap": true, 00:08:55.331 "flush": true, 00:08:55.331 "reset": true, 00:08:55.331 "nvme_admin": false, 00:08:55.331 "nvme_io": false, 00:08:55.331 "nvme_io_md": false, 00:08:55.331 "write_zeroes": true, 00:08:55.331 "zcopy": true, 00:08:55.331 "get_zone_info": false, 00:08:55.331 "zone_management": false, 00:08:55.331 "zone_append": false, 00:08:55.331 "compare": false, 00:08:55.331 "compare_and_write": false, 00:08:55.331 "abort": true, 00:08:55.331 "seek_hole": false, 00:08:55.331 "seek_data": false, 00:08:55.331 "copy": true, 00:08:55.331 "nvme_iov_md": false 00:08:55.332 }, 00:08:55.332 "memory_domains": [ 00:08:55.332 { 00:08:55.332 "dma_device_id": "system", 00:08:55.332 "dma_device_type": 1 00:08:55.332 }, 00:08:55.332 { 00:08:55.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.332 "dma_device_type": 2 00:08:55.332 } 00:08:55.332 ], 00:08:55.332 "driver_specific": {} 00:08:55.332 } 00:08:55.332 ] 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.332 "name": "Existed_Raid", 00:08:55.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:55.332 "strip_size_kb": 0, 00:08:55.332 "state": "configuring", 00:08:55.332 "raid_level": "raid1", 00:08:55.332 "superblock": false, 00:08:55.332 "num_base_bdevs": 3, 00:08:55.332 "num_base_bdevs_discovered": 2, 00:08:55.332 "num_base_bdevs_operational": 3, 00:08:55.332 "base_bdevs_list": [ 00:08:55.332 { 00:08:55.332 "name": "BaseBdev1", 00:08:55.332 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:55.332 "is_configured": true, 00:08:55.332 "data_offset": 0, 00:08:55.332 "data_size": 65536 00:08:55.332 }, 00:08:55.332 { 00:08:55.332 "name": null, 00:08:55.332 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:55.332 "is_configured": false, 00:08:55.332 "data_offset": 0, 00:08:55.332 "data_size": 65536 00:08:55.332 }, 00:08:55.332 { 00:08:55.332 "name": "BaseBdev3", 00:08:55.332 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:55.332 "is_configured": true, 00:08:55.332 "data_offset": 0, 00:08:55.332 "data_size": 65536 00:08:55.332 } 00:08:55.332 ] 00:08:55.332 }' 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.332 14:07:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.900 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:55.900 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.900 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.900 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.900 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.901 [2024-09-30 14:08:00.347059] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.901 "name": "Existed_Raid", 00:08:55.901 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:55.901 "strip_size_kb": 0, 00:08:55.901 "state": "configuring", 00:08:55.901 "raid_level": "raid1", 00:08:55.901 "superblock": false, 00:08:55.901 "num_base_bdevs": 3, 00:08:55.901 "num_base_bdevs_discovered": 1, 00:08:55.901 "num_base_bdevs_operational": 3, 00:08:55.901 "base_bdevs_list": [ 00:08:55.901 { 00:08:55.901 "name": "BaseBdev1", 00:08:55.901 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:55.901 "is_configured": true, 00:08:55.901 "data_offset": 0, 00:08:55.901 "data_size": 65536 00:08:55.901 }, 00:08:55.901 { 00:08:55.901 "name": null, 00:08:55.901 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:55.901 "is_configured": false, 00:08:55.901 "data_offset": 0, 00:08:55.901 "data_size": 65536 00:08:55.901 }, 00:08:55.901 { 00:08:55.901 "name": null, 00:08:55.901 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:55.901 "is_configured": false, 00:08:55.901 "data_offset": 0, 00:08:55.901 "data_size": 65536 00:08:55.901 } 00:08:55.901 ] 00:08:55.901 }' 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.901 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:56.160 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.419 [2024-09-30 14:08:00.822320] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.419 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.419 "name": "Existed_Raid", 00:08:56.419 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.419 "strip_size_kb": 0, 00:08:56.419 "state": "configuring", 00:08:56.419 "raid_level": "raid1", 00:08:56.419 "superblock": false, 00:08:56.419 "num_base_bdevs": 3, 00:08:56.419 "num_base_bdevs_discovered": 2, 00:08:56.419 "num_base_bdevs_operational": 3, 00:08:56.419 "base_bdevs_list": [ 00:08:56.419 { 00:08:56.420 "name": "BaseBdev1", 00:08:56.420 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:56.420 "is_configured": true, 00:08:56.420 "data_offset": 0, 00:08:56.420 "data_size": 65536 00:08:56.420 }, 00:08:56.420 { 00:08:56.420 "name": null, 00:08:56.420 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:56.420 "is_configured": false, 00:08:56.420 "data_offset": 0, 00:08:56.420 "data_size": 65536 00:08:56.420 }, 00:08:56.420 { 00:08:56.420 "name": "BaseBdev3", 00:08:56.420 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:56.420 "is_configured": true, 00:08:56.420 "data_offset": 0, 00:08:56.420 "data_size": 65536 00:08:56.420 } 00:08:56.420 ] 00:08:56.420 }' 00:08:56.420 14:08:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.420 14:08:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.679 [2024-09-30 14:08:01.305549] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.679 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.938 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.938 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.938 "name": "Existed_Raid", 00:08:56.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.938 "strip_size_kb": 0, 00:08:56.938 "state": "configuring", 00:08:56.938 "raid_level": "raid1", 00:08:56.938 "superblock": false, 00:08:56.938 "num_base_bdevs": 3, 00:08:56.938 "num_base_bdevs_discovered": 1, 00:08:56.938 "num_base_bdevs_operational": 3, 00:08:56.938 "base_bdevs_list": [ 00:08:56.938 { 00:08:56.938 "name": null, 00:08:56.938 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:56.939 "is_configured": false, 00:08:56.939 "data_offset": 0, 00:08:56.939 "data_size": 65536 00:08:56.939 }, 00:08:56.939 { 00:08:56.939 "name": null, 00:08:56.939 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:56.939 "is_configured": false, 00:08:56.939 "data_offset": 0, 00:08:56.939 "data_size": 65536 00:08:56.939 }, 00:08:56.939 { 00:08:56.939 "name": "BaseBdev3", 00:08:56.939 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:56.939 "is_configured": true, 00:08:56.939 "data_offset": 0, 00:08:56.939 "data_size": 65536 00:08:56.939 } 00:08:56.939 ] 00:08:56.939 }' 00:08:56.939 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.939 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.198 [2024-09-30 14:08:01.803122] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.198 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.457 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.457 "name": "Existed_Raid", 00:08:57.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.457 "strip_size_kb": 0, 00:08:57.457 "state": "configuring", 00:08:57.457 "raid_level": "raid1", 00:08:57.457 "superblock": false, 00:08:57.457 "num_base_bdevs": 3, 00:08:57.457 "num_base_bdevs_discovered": 2, 00:08:57.457 "num_base_bdevs_operational": 3, 00:08:57.457 "base_bdevs_list": [ 00:08:57.457 { 00:08:57.457 "name": null, 00:08:57.457 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:57.457 "is_configured": false, 00:08:57.457 "data_offset": 0, 00:08:57.457 "data_size": 65536 00:08:57.457 }, 00:08:57.457 { 00:08:57.457 "name": "BaseBdev2", 00:08:57.457 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:57.457 "is_configured": true, 00:08:57.457 "data_offset": 0, 00:08:57.457 "data_size": 65536 00:08:57.457 }, 00:08:57.457 { 00:08:57.457 "name": "BaseBdev3", 00:08:57.457 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:57.457 "is_configured": true, 00:08:57.457 "data_offset": 0, 00:08:57.457 "data_size": 65536 00:08:57.457 } 00:08:57.457 ] 00:08:57.457 }' 00:08:57.457 14:08:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.457 14:08:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u acf85c78-3dbd-4e33-acac-65e1e93fdf7b 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.717 [2024-09-30 14:08:02.368974] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:57.717 [2024-09-30 14:08:02.369079] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:08:57.717 [2024-09-30 14:08:02.369126] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:57.717 [2024-09-30 14:08:02.369406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:08:57.717 [2024-09-30 14:08:02.369580] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:08:57.717 [2024-09-30 14:08:02.369621] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:08:57.717 [2024-09-30 14:08:02.369826] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.717 NewBaseBdev 00:08:57.717 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.976 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.977 [ 00:08:57.977 { 00:08:57.977 "name": "NewBaseBdev", 00:08:57.977 "aliases": [ 00:08:57.977 "acf85c78-3dbd-4e33-acac-65e1e93fdf7b" 00:08:57.977 ], 00:08:57.977 "product_name": "Malloc disk", 00:08:57.977 "block_size": 512, 00:08:57.977 "num_blocks": 65536, 00:08:57.977 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:57.977 "assigned_rate_limits": { 00:08:57.977 "rw_ios_per_sec": 0, 00:08:57.977 "rw_mbytes_per_sec": 0, 00:08:57.977 "r_mbytes_per_sec": 0, 00:08:57.977 "w_mbytes_per_sec": 0 00:08:57.977 }, 00:08:57.977 "claimed": true, 00:08:57.977 "claim_type": "exclusive_write", 00:08:57.977 "zoned": false, 00:08:57.977 "supported_io_types": { 00:08:57.977 "read": true, 00:08:57.977 "write": true, 00:08:57.977 "unmap": true, 00:08:57.977 "flush": true, 00:08:57.977 "reset": true, 00:08:57.977 "nvme_admin": false, 00:08:57.977 "nvme_io": false, 00:08:57.977 "nvme_io_md": false, 00:08:57.977 "write_zeroes": true, 00:08:57.977 "zcopy": true, 00:08:57.977 "get_zone_info": false, 00:08:57.977 "zone_management": false, 00:08:57.977 "zone_append": false, 00:08:57.977 "compare": false, 00:08:57.977 "compare_and_write": false, 00:08:57.977 "abort": true, 00:08:57.977 "seek_hole": false, 00:08:57.977 "seek_data": false, 00:08:57.977 "copy": true, 00:08:57.977 "nvme_iov_md": false 00:08:57.977 }, 00:08:57.977 "memory_domains": [ 00:08:57.977 { 00:08:57.977 "dma_device_id": "system", 00:08:57.977 "dma_device_type": 1 00:08:57.977 }, 00:08:57.977 { 00:08:57.977 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.977 "dma_device_type": 2 00:08:57.977 } 00:08:57.977 ], 00:08:57.977 "driver_specific": {} 00:08:57.977 } 00:08:57.977 ] 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.977 "name": "Existed_Raid", 00:08:57.977 "uuid": "eb64d4ba-9481-40b2-b4ed-5b128972eef0", 00:08:57.977 "strip_size_kb": 0, 00:08:57.977 "state": "online", 00:08:57.977 "raid_level": "raid1", 00:08:57.977 "superblock": false, 00:08:57.977 "num_base_bdevs": 3, 00:08:57.977 "num_base_bdevs_discovered": 3, 00:08:57.977 "num_base_bdevs_operational": 3, 00:08:57.977 "base_bdevs_list": [ 00:08:57.977 { 00:08:57.977 "name": "NewBaseBdev", 00:08:57.977 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:57.977 "is_configured": true, 00:08:57.977 "data_offset": 0, 00:08:57.977 "data_size": 65536 00:08:57.977 }, 00:08:57.977 { 00:08:57.977 "name": "BaseBdev2", 00:08:57.977 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:57.977 "is_configured": true, 00:08:57.977 "data_offset": 0, 00:08:57.977 "data_size": 65536 00:08:57.977 }, 00:08:57.977 { 00:08:57.977 "name": "BaseBdev3", 00:08:57.977 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:57.977 "is_configured": true, 00:08:57.977 "data_offset": 0, 00:08:57.977 "data_size": 65536 00:08:57.977 } 00:08:57.977 ] 00:08:57.977 }' 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.977 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.239 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:58.239 [2024-09-30 14:08:02.892403] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.529 14:08:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.529 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.529 "name": "Existed_Raid", 00:08:58.529 "aliases": [ 00:08:58.529 "eb64d4ba-9481-40b2-b4ed-5b128972eef0" 00:08:58.529 ], 00:08:58.529 "product_name": "Raid Volume", 00:08:58.529 "block_size": 512, 00:08:58.529 "num_blocks": 65536, 00:08:58.529 "uuid": "eb64d4ba-9481-40b2-b4ed-5b128972eef0", 00:08:58.529 "assigned_rate_limits": { 00:08:58.529 "rw_ios_per_sec": 0, 00:08:58.529 "rw_mbytes_per_sec": 0, 00:08:58.529 "r_mbytes_per_sec": 0, 00:08:58.529 "w_mbytes_per_sec": 0 00:08:58.529 }, 00:08:58.529 "claimed": false, 00:08:58.529 "zoned": false, 00:08:58.529 "supported_io_types": { 00:08:58.529 "read": true, 00:08:58.529 "write": true, 00:08:58.529 "unmap": false, 00:08:58.529 "flush": false, 00:08:58.529 "reset": true, 00:08:58.529 "nvme_admin": false, 00:08:58.529 "nvme_io": false, 00:08:58.529 "nvme_io_md": false, 00:08:58.529 "write_zeroes": true, 00:08:58.529 "zcopy": false, 00:08:58.529 "get_zone_info": false, 00:08:58.529 "zone_management": false, 00:08:58.529 "zone_append": false, 00:08:58.529 "compare": false, 00:08:58.529 "compare_and_write": false, 00:08:58.529 "abort": false, 00:08:58.529 "seek_hole": false, 00:08:58.529 "seek_data": false, 00:08:58.529 "copy": false, 00:08:58.529 "nvme_iov_md": false 00:08:58.529 }, 00:08:58.529 "memory_domains": [ 00:08:58.529 { 00:08:58.529 "dma_device_id": "system", 00:08:58.529 "dma_device_type": 1 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.529 "dma_device_type": 2 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "dma_device_id": "system", 00:08:58.529 "dma_device_type": 1 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.529 "dma_device_type": 2 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "dma_device_id": "system", 00:08:58.529 "dma_device_type": 1 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.529 "dma_device_type": 2 00:08:58.529 } 00:08:58.529 ], 00:08:58.529 "driver_specific": { 00:08:58.529 "raid": { 00:08:58.529 "uuid": "eb64d4ba-9481-40b2-b4ed-5b128972eef0", 00:08:58.529 "strip_size_kb": 0, 00:08:58.529 "state": "online", 00:08:58.529 "raid_level": "raid1", 00:08:58.529 "superblock": false, 00:08:58.529 "num_base_bdevs": 3, 00:08:58.529 "num_base_bdevs_discovered": 3, 00:08:58.529 "num_base_bdevs_operational": 3, 00:08:58.529 "base_bdevs_list": [ 00:08:58.529 { 00:08:58.529 "name": "NewBaseBdev", 00:08:58.529 "uuid": "acf85c78-3dbd-4e33-acac-65e1e93fdf7b", 00:08:58.529 "is_configured": true, 00:08:58.529 "data_offset": 0, 00:08:58.529 "data_size": 65536 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "name": "BaseBdev2", 00:08:58.529 "uuid": "8b673c82-0166-493a-80c8-81f8a58f1526", 00:08:58.529 "is_configured": true, 00:08:58.529 "data_offset": 0, 00:08:58.529 "data_size": 65536 00:08:58.529 }, 00:08:58.529 { 00:08:58.529 "name": "BaseBdev3", 00:08:58.529 "uuid": "cac8d6bd-5496-4175-a4dd-f36c5f814cd2", 00:08:58.529 "is_configured": true, 00:08:58.529 "data_offset": 0, 00:08:58.529 "data_size": 65536 00:08:58.529 } 00:08:58.529 ] 00:08:58.529 } 00:08:58.529 } 00:08:58.529 }' 00:08:58.529 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.529 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:58.529 BaseBdev2 00:08:58.529 BaseBdev3' 00:08:58.529 14:08:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.529 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.529 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.529 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.530 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.530 [2024-09-30 14:08:03.167630] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:58.530 [2024-09-30 14:08:03.167701] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.530 [2024-09-30 14:08:03.167788] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.530 [2024-09-30 14:08:03.168083] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:58.530 [2024-09-30 14:08:03.168138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 79484 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 79484 ']' 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 79484 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79484 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79484' 00:08:58.797 killing process with pid 79484 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 79484 00:08:58.797 [2024-09-30 14:08:03.216140] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:58.797 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 79484 00:08:58.797 [2024-09-30 14:08:03.246670] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:59.057 00:08:59.057 real 0m8.871s 00:08:59.057 user 0m15.160s 00:08:59.057 sys 0m1.780s 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:59.057 ************************************ 00:08:59.057 END TEST raid_state_function_test 00:08:59.057 ************************************ 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.057 14:08:03 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:08:59.057 14:08:03 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:59.057 14:08:03 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:59.057 14:08:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:59.057 ************************************ 00:08:59.057 START TEST raid_state_function_test_sb 00:08:59.057 ************************************ 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 true 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:59.057 Process raid pid: 80094 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80094 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80094' 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80094 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 80094 ']' 00:08:59.057 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:59.057 14:08:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.057 [2024-09-30 14:08:03.657502] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:08:59.057 [2024-09-30 14:08:03.657638] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:59.317 [2024-09-30 14:08:03.790152] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:08:59.317 [2024-09-30 14:08:03.818647] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:59.317 [2024-09-30 14:08:03.863567] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:59.317 [2024-09-30 14:08:03.904978] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:59.317 [2024-09-30 14:08:03.905012] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.886 [2024-09-30 14:08:04.497923] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:59.886 [2024-09-30 14:08:04.497969] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:59.886 [2024-09-30 14:08:04.497980] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:59.886 [2024-09-30 14:08:04.497988] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:59.886 [2024-09-30 14:08:04.497998] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:59.886 [2024-09-30 14:08:04.498004] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.886 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.146 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.146 "name": "Existed_Raid", 00:09:00.146 "uuid": "f4e6b32c-6d14-401e-91b3-e3bf0d6a6f7d", 00:09:00.146 "strip_size_kb": 0, 00:09:00.146 "state": "configuring", 00:09:00.146 "raid_level": "raid1", 00:09:00.146 "superblock": true, 00:09:00.146 "num_base_bdevs": 3, 00:09:00.146 "num_base_bdevs_discovered": 0, 00:09:00.146 "num_base_bdevs_operational": 3, 00:09:00.146 "base_bdevs_list": [ 00:09:00.146 { 00:09:00.146 "name": "BaseBdev1", 00:09:00.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.146 "is_configured": false, 00:09:00.146 "data_offset": 0, 00:09:00.146 "data_size": 0 00:09:00.146 }, 00:09:00.146 { 00:09:00.146 "name": "BaseBdev2", 00:09:00.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.146 "is_configured": false, 00:09:00.146 "data_offset": 0, 00:09:00.146 "data_size": 0 00:09:00.146 }, 00:09:00.146 { 00:09:00.146 "name": "BaseBdev3", 00:09:00.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.146 "is_configured": false, 00:09:00.146 "data_offset": 0, 00:09:00.146 "data_size": 0 00:09:00.146 } 00:09:00.146 ] 00:09:00.146 }' 00:09:00.146 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.146 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 [2024-09-30 14:08:04.865196] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.406 [2024-09-30 14:08:04.865283] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 [2024-09-30 14:08:04.877193] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:00.406 [2024-09-30 14:08:04.877269] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:00.406 [2024-09-30 14:08:04.877304] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.406 [2024-09-30 14:08:04.877323] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.406 [2024-09-30 14:08:04.877341] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:00.406 [2024-09-30 14:08:04.877358] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 [2024-09-30 14:08:04.897739] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.406 BaseBdev1 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 [ 00:09:00.406 { 00:09:00.406 "name": "BaseBdev1", 00:09:00.406 "aliases": [ 00:09:00.406 "a1d16841-2a56-428a-96fa-221e6336d9a2" 00:09:00.406 ], 00:09:00.406 "product_name": "Malloc disk", 00:09:00.406 "block_size": 512, 00:09:00.406 "num_blocks": 65536, 00:09:00.406 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:00.406 "assigned_rate_limits": { 00:09:00.406 "rw_ios_per_sec": 0, 00:09:00.406 "rw_mbytes_per_sec": 0, 00:09:00.406 "r_mbytes_per_sec": 0, 00:09:00.406 "w_mbytes_per_sec": 0 00:09:00.406 }, 00:09:00.406 "claimed": true, 00:09:00.406 "claim_type": "exclusive_write", 00:09:00.406 "zoned": false, 00:09:00.406 "supported_io_types": { 00:09:00.406 "read": true, 00:09:00.406 "write": true, 00:09:00.406 "unmap": true, 00:09:00.406 "flush": true, 00:09:00.406 "reset": true, 00:09:00.406 "nvme_admin": false, 00:09:00.406 "nvme_io": false, 00:09:00.406 "nvme_io_md": false, 00:09:00.406 "write_zeroes": true, 00:09:00.406 "zcopy": true, 00:09:00.406 "get_zone_info": false, 00:09:00.406 "zone_management": false, 00:09:00.406 "zone_append": false, 00:09:00.406 "compare": false, 00:09:00.406 "compare_and_write": false, 00:09:00.406 "abort": true, 00:09:00.406 "seek_hole": false, 00:09:00.406 "seek_data": false, 00:09:00.406 "copy": true, 00:09:00.406 "nvme_iov_md": false 00:09:00.406 }, 00:09:00.406 "memory_domains": [ 00:09:00.406 { 00:09:00.406 "dma_device_id": "system", 00:09:00.406 "dma_device_type": 1 00:09:00.406 }, 00:09:00.406 { 00:09:00.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.406 "dma_device_type": 2 00:09:00.406 } 00:09:00.406 ], 00:09:00.406 "driver_specific": {} 00:09:00.406 } 00:09:00.406 ] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.406 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.406 "name": "Existed_Raid", 00:09:00.406 "uuid": "91f97b03-a5e5-410e-abd4-525f070de238", 00:09:00.406 "strip_size_kb": 0, 00:09:00.406 "state": "configuring", 00:09:00.406 "raid_level": "raid1", 00:09:00.406 "superblock": true, 00:09:00.406 "num_base_bdevs": 3, 00:09:00.406 "num_base_bdevs_discovered": 1, 00:09:00.406 "num_base_bdevs_operational": 3, 00:09:00.406 "base_bdevs_list": [ 00:09:00.406 { 00:09:00.406 "name": "BaseBdev1", 00:09:00.406 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:00.406 "is_configured": true, 00:09:00.406 "data_offset": 2048, 00:09:00.406 "data_size": 63488 00:09:00.406 }, 00:09:00.406 { 00:09:00.406 "name": "BaseBdev2", 00:09:00.406 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.406 "is_configured": false, 00:09:00.406 "data_offset": 0, 00:09:00.406 "data_size": 0 00:09:00.406 }, 00:09:00.406 { 00:09:00.407 "name": "BaseBdev3", 00:09:00.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.407 "is_configured": false, 00:09:00.407 "data_offset": 0, 00:09:00.407 "data_size": 0 00:09:00.407 } 00:09:00.407 ] 00:09:00.407 }' 00:09:00.407 14:08:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.407 14:08:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.975 [2024-09-30 14:08:05.400897] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.975 [2024-09-30 14:08:05.400995] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.975 [2024-09-30 14:08:05.412900] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.975 [2024-09-30 14:08:05.414732] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.975 [2024-09-30 14:08:05.414768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.975 [2024-09-30 14:08:05.414780] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:00.975 [2024-09-30 14:08:05.414787] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.975 "name": "Existed_Raid", 00:09:00.975 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:00.975 "strip_size_kb": 0, 00:09:00.975 "state": "configuring", 00:09:00.975 "raid_level": "raid1", 00:09:00.975 "superblock": true, 00:09:00.975 "num_base_bdevs": 3, 00:09:00.975 "num_base_bdevs_discovered": 1, 00:09:00.975 "num_base_bdevs_operational": 3, 00:09:00.975 "base_bdevs_list": [ 00:09:00.975 { 00:09:00.975 "name": "BaseBdev1", 00:09:00.975 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:00.975 "is_configured": true, 00:09:00.975 "data_offset": 2048, 00:09:00.975 "data_size": 63488 00:09:00.975 }, 00:09:00.975 { 00:09:00.975 "name": "BaseBdev2", 00:09:00.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.975 "is_configured": false, 00:09:00.975 "data_offset": 0, 00:09:00.975 "data_size": 0 00:09:00.975 }, 00:09:00.975 { 00:09:00.975 "name": "BaseBdev3", 00:09:00.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.975 "is_configured": false, 00:09:00.975 "data_offset": 0, 00:09:00.975 "data_size": 0 00:09:00.975 } 00:09:00.975 ] 00:09:00.975 }' 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.975 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.235 [2024-09-30 14:08:05.868906] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:01.235 BaseBdev2 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.235 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.493 [ 00:09:01.493 { 00:09:01.493 "name": "BaseBdev2", 00:09:01.493 "aliases": [ 00:09:01.494 "5c0011e0-f810-4db1-8eac-77a1a1832e5a" 00:09:01.494 ], 00:09:01.494 "product_name": "Malloc disk", 00:09:01.494 "block_size": 512, 00:09:01.494 "num_blocks": 65536, 00:09:01.494 "uuid": "5c0011e0-f810-4db1-8eac-77a1a1832e5a", 00:09:01.494 "assigned_rate_limits": { 00:09:01.494 "rw_ios_per_sec": 0, 00:09:01.494 "rw_mbytes_per_sec": 0, 00:09:01.494 "r_mbytes_per_sec": 0, 00:09:01.494 "w_mbytes_per_sec": 0 00:09:01.494 }, 00:09:01.494 "claimed": true, 00:09:01.494 "claim_type": "exclusive_write", 00:09:01.494 "zoned": false, 00:09:01.494 "supported_io_types": { 00:09:01.494 "read": true, 00:09:01.494 "write": true, 00:09:01.494 "unmap": true, 00:09:01.494 "flush": true, 00:09:01.494 "reset": true, 00:09:01.494 "nvme_admin": false, 00:09:01.494 "nvme_io": false, 00:09:01.494 "nvme_io_md": false, 00:09:01.494 "write_zeroes": true, 00:09:01.494 "zcopy": true, 00:09:01.494 "get_zone_info": false, 00:09:01.494 "zone_management": false, 00:09:01.494 "zone_append": false, 00:09:01.494 "compare": false, 00:09:01.494 "compare_and_write": false, 00:09:01.494 "abort": true, 00:09:01.494 "seek_hole": false, 00:09:01.494 "seek_data": false, 00:09:01.494 "copy": true, 00:09:01.494 "nvme_iov_md": false 00:09:01.494 }, 00:09:01.494 "memory_domains": [ 00:09:01.494 { 00:09:01.494 "dma_device_id": "system", 00:09:01.494 "dma_device_type": 1 00:09:01.494 }, 00:09:01.494 { 00:09:01.494 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.494 "dma_device_type": 2 00:09:01.494 } 00:09:01.494 ], 00:09:01.494 "driver_specific": {} 00:09:01.494 } 00:09:01.494 ] 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.494 "name": "Existed_Raid", 00:09:01.494 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:01.494 "strip_size_kb": 0, 00:09:01.494 "state": "configuring", 00:09:01.494 "raid_level": "raid1", 00:09:01.494 "superblock": true, 00:09:01.494 "num_base_bdevs": 3, 00:09:01.494 "num_base_bdevs_discovered": 2, 00:09:01.494 "num_base_bdevs_operational": 3, 00:09:01.494 "base_bdevs_list": [ 00:09:01.494 { 00:09:01.494 "name": "BaseBdev1", 00:09:01.494 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:01.494 "is_configured": true, 00:09:01.494 "data_offset": 2048, 00:09:01.494 "data_size": 63488 00:09:01.494 }, 00:09:01.494 { 00:09:01.494 "name": "BaseBdev2", 00:09:01.494 "uuid": "5c0011e0-f810-4db1-8eac-77a1a1832e5a", 00:09:01.494 "is_configured": true, 00:09:01.494 "data_offset": 2048, 00:09:01.494 "data_size": 63488 00:09:01.494 }, 00:09:01.494 { 00:09:01.494 "name": "BaseBdev3", 00:09:01.494 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.494 "is_configured": false, 00:09:01.494 "data_offset": 0, 00:09:01.494 "data_size": 0 00:09:01.494 } 00:09:01.494 ] 00:09:01.494 }' 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.494 14:08:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.768 [2024-09-30 14:08:06.354853] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:01.768 [2024-09-30 14:08:06.355114] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:01.768 [2024-09-30 14:08:06.355153] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:01.768 BaseBdev3 00:09:01.768 [2024-09-30 14:08:06.355451] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:01.768 [2024-09-30 14:08:06.355609] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:01.768 [2024-09-30 14:08:06.355629] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:01.768 [2024-09-30 14:08:06.355759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.768 [ 00:09:01.768 { 00:09:01.768 "name": "BaseBdev3", 00:09:01.768 "aliases": [ 00:09:01.768 "00fb4796-dbd5-4fa4-af09-4b0a631ca558" 00:09:01.768 ], 00:09:01.768 "product_name": "Malloc disk", 00:09:01.768 "block_size": 512, 00:09:01.768 "num_blocks": 65536, 00:09:01.768 "uuid": "00fb4796-dbd5-4fa4-af09-4b0a631ca558", 00:09:01.768 "assigned_rate_limits": { 00:09:01.768 "rw_ios_per_sec": 0, 00:09:01.768 "rw_mbytes_per_sec": 0, 00:09:01.768 "r_mbytes_per_sec": 0, 00:09:01.768 "w_mbytes_per_sec": 0 00:09:01.768 }, 00:09:01.768 "claimed": true, 00:09:01.768 "claim_type": "exclusive_write", 00:09:01.768 "zoned": false, 00:09:01.768 "supported_io_types": { 00:09:01.768 "read": true, 00:09:01.768 "write": true, 00:09:01.768 "unmap": true, 00:09:01.768 "flush": true, 00:09:01.768 "reset": true, 00:09:01.768 "nvme_admin": false, 00:09:01.768 "nvme_io": false, 00:09:01.768 "nvme_io_md": false, 00:09:01.768 "write_zeroes": true, 00:09:01.768 "zcopy": true, 00:09:01.768 "get_zone_info": false, 00:09:01.768 "zone_management": false, 00:09:01.768 "zone_append": false, 00:09:01.768 "compare": false, 00:09:01.768 "compare_and_write": false, 00:09:01.768 "abort": true, 00:09:01.768 "seek_hole": false, 00:09:01.768 "seek_data": false, 00:09:01.768 "copy": true, 00:09:01.768 "nvme_iov_md": false 00:09:01.768 }, 00:09:01.768 "memory_domains": [ 00:09:01.768 { 00:09:01.768 "dma_device_id": "system", 00:09:01.768 "dma_device_type": 1 00:09:01.768 }, 00:09:01.768 { 00:09:01.768 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.768 "dma_device_type": 2 00:09:01.768 } 00:09:01.768 ], 00:09:01.768 "driver_specific": {} 00:09:01.768 } 00:09:01.768 ] 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.768 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.027 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.027 "name": "Existed_Raid", 00:09:02.027 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:02.027 "strip_size_kb": 0, 00:09:02.027 "state": "online", 00:09:02.027 "raid_level": "raid1", 00:09:02.027 "superblock": true, 00:09:02.027 "num_base_bdevs": 3, 00:09:02.027 "num_base_bdevs_discovered": 3, 00:09:02.027 "num_base_bdevs_operational": 3, 00:09:02.027 "base_bdevs_list": [ 00:09:02.027 { 00:09:02.027 "name": "BaseBdev1", 00:09:02.027 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:02.027 "is_configured": true, 00:09:02.027 "data_offset": 2048, 00:09:02.027 "data_size": 63488 00:09:02.027 }, 00:09:02.027 { 00:09:02.027 "name": "BaseBdev2", 00:09:02.027 "uuid": "5c0011e0-f810-4db1-8eac-77a1a1832e5a", 00:09:02.027 "is_configured": true, 00:09:02.027 "data_offset": 2048, 00:09:02.027 "data_size": 63488 00:09:02.027 }, 00:09:02.027 { 00:09:02.027 "name": "BaseBdev3", 00:09:02.027 "uuid": "00fb4796-dbd5-4fa4-af09-4b0a631ca558", 00:09:02.027 "is_configured": true, 00:09:02.027 "data_offset": 2048, 00:09:02.027 "data_size": 63488 00:09:02.027 } 00:09:02.027 ] 00:09:02.027 }' 00:09:02.027 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.027 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.286 [2024-09-30 14:08:06.810398] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:02.286 "name": "Existed_Raid", 00:09:02.286 "aliases": [ 00:09:02.286 "ededc517-d6ef-4e35-974f-9e2f993dfd72" 00:09:02.286 ], 00:09:02.286 "product_name": "Raid Volume", 00:09:02.286 "block_size": 512, 00:09:02.286 "num_blocks": 63488, 00:09:02.286 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:02.286 "assigned_rate_limits": { 00:09:02.286 "rw_ios_per_sec": 0, 00:09:02.286 "rw_mbytes_per_sec": 0, 00:09:02.286 "r_mbytes_per_sec": 0, 00:09:02.286 "w_mbytes_per_sec": 0 00:09:02.286 }, 00:09:02.286 "claimed": false, 00:09:02.286 "zoned": false, 00:09:02.286 "supported_io_types": { 00:09:02.286 "read": true, 00:09:02.286 "write": true, 00:09:02.286 "unmap": false, 00:09:02.286 "flush": false, 00:09:02.286 "reset": true, 00:09:02.286 "nvme_admin": false, 00:09:02.286 "nvme_io": false, 00:09:02.286 "nvme_io_md": false, 00:09:02.286 "write_zeroes": true, 00:09:02.286 "zcopy": false, 00:09:02.286 "get_zone_info": false, 00:09:02.286 "zone_management": false, 00:09:02.286 "zone_append": false, 00:09:02.286 "compare": false, 00:09:02.286 "compare_and_write": false, 00:09:02.286 "abort": false, 00:09:02.286 "seek_hole": false, 00:09:02.286 "seek_data": false, 00:09:02.286 "copy": false, 00:09:02.286 "nvme_iov_md": false 00:09:02.286 }, 00:09:02.286 "memory_domains": [ 00:09:02.286 { 00:09:02.286 "dma_device_id": "system", 00:09:02.286 "dma_device_type": 1 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.286 "dma_device_type": 2 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "dma_device_id": "system", 00:09:02.286 "dma_device_type": 1 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.286 "dma_device_type": 2 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "dma_device_id": "system", 00:09:02.286 "dma_device_type": 1 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.286 "dma_device_type": 2 00:09:02.286 } 00:09:02.286 ], 00:09:02.286 "driver_specific": { 00:09:02.286 "raid": { 00:09:02.286 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:02.286 "strip_size_kb": 0, 00:09:02.286 "state": "online", 00:09:02.286 "raid_level": "raid1", 00:09:02.286 "superblock": true, 00:09:02.286 "num_base_bdevs": 3, 00:09:02.286 "num_base_bdevs_discovered": 3, 00:09:02.286 "num_base_bdevs_operational": 3, 00:09:02.286 "base_bdevs_list": [ 00:09:02.286 { 00:09:02.286 "name": "BaseBdev1", 00:09:02.286 "uuid": "a1d16841-2a56-428a-96fa-221e6336d9a2", 00:09:02.286 "is_configured": true, 00:09:02.286 "data_offset": 2048, 00:09:02.286 "data_size": 63488 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "name": "BaseBdev2", 00:09:02.286 "uuid": "5c0011e0-f810-4db1-8eac-77a1a1832e5a", 00:09:02.286 "is_configured": true, 00:09:02.286 "data_offset": 2048, 00:09:02.286 "data_size": 63488 00:09:02.286 }, 00:09:02.286 { 00:09:02.286 "name": "BaseBdev3", 00:09:02.286 "uuid": "00fb4796-dbd5-4fa4-af09-4b0a631ca558", 00:09:02.286 "is_configured": true, 00:09:02.286 "data_offset": 2048, 00:09:02.286 "data_size": 63488 00:09:02.286 } 00:09:02.286 ] 00:09:02.286 } 00:09:02.286 } 00:09:02.286 }' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:02.286 BaseBdev2 00:09:02.286 BaseBdev3' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.286 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.545 14:08:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.545 [2024-09-30 14:08:07.045772] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.545 "name": "Existed_Raid", 00:09:02.545 "uuid": "ededc517-d6ef-4e35-974f-9e2f993dfd72", 00:09:02.545 "strip_size_kb": 0, 00:09:02.545 "state": "online", 00:09:02.545 "raid_level": "raid1", 00:09:02.545 "superblock": true, 00:09:02.545 "num_base_bdevs": 3, 00:09:02.545 "num_base_bdevs_discovered": 2, 00:09:02.545 "num_base_bdevs_operational": 2, 00:09:02.545 "base_bdevs_list": [ 00:09:02.545 { 00:09:02.545 "name": null, 00:09:02.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.545 "is_configured": false, 00:09:02.545 "data_offset": 0, 00:09:02.545 "data_size": 63488 00:09:02.545 }, 00:09:02.545 { 00:09:02.545 "name": "BaseBdev2", 00:09:02.545 "uuid": "5c0011e0-f810-4db1-8eac-77a1a1832e5a", 00:09:02.545 "is_configured": true, 00:09:02.545 "data_offset": 2048, 00:09:02.545 "data_size": 63488 00:09:02.545 }, 00:09:02.545 { 00:09:02.545 "name": "BaseBdev3", 00:09:02.545 "uuid": "00fb4796-dbd5-4fa4-af09-4b0a631ca558", 00:09:02.545 "is_configured": true, 00:09:02.545 "data_offset": 2048, 00:09:02.545 "data_size": 63488 00:09:02.545 } 00:09:02.545 ] 00:09:02.545 }' 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.545 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.114 [2024-09-30 14:08:07.568099] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.114 [2024-09-30 14:08:07.639059] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:03.114 [2024-09-30 14:08:07.639203] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:03.114 [2024-09-30 14:08:07.650608] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:03.114 [2024-09-30 14:08:07.650746] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:03.114 [2024-09-30 14:08:07.650787] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.114 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.115 BaseBdev2 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.115 [ 00:09:03.115 { 00:09:03.115 "name": "BaseBdev2", 00:09:03.115 "aliases": [ 00:09:03.115 "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1" 00:09:03.115 ], 00:09:03.115 "product_name": "Malloc disk", 00:09:03.115 "block_size": 512, 00:09:03.115 "num_blocks": 65536, 00:09:03.115 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:03.115 "assigned_rate_limits": { 00:09:03.115 "rw_ios_per_sec": 0, 00:09:03.115 "rw_mbytes_per_sec": 0, 00:09:03.115 "r_mbytes_per_sec": 0, 00:09:03.115 "w_mbytes_per_sec": 0 00:09:03.115 }, 00:09:03.115 "claimed": false, 00:09:03.115 "zoned": false, 00:09:03.115 "supported_io_types": { 00:09:03.115 "read": true, 00:09:03.115 "write": true, 00:09:03.115 "unmap": true, 00:09:03.115 "flush": true, 00:09:03.115 "reset": true, 00:09:03.115 "nvme_admin": false, 00:09:03.115 "nvme_io": false, 00:09:03.115 "nvme_io_md": false, 00:09:03.115 "write_zeroes": true, 00:09:03.115 "zcopy": true, 00:09:03.115 "get_zone_info": false, 00:09:03.115 "zone_management": false, 00:09:03.115 "zone_append": false, 00:09:03.115 "compare": false, 00:09:03.115 "compare_and_write": false, 00:09:03.115 "abort": true, 00:09:03.115 "seek_hole": false, 00:09:03.115 "seek_data": false, 00:09:03.115 "copy": true, 00:09:03.115 "nvme_iov_md": false 00:09:03.115 }, 00:09:03.115 "memory_domains": [ 00:09:03.115 { 00:09:03.115 "dma_device_id": "system", 00:09:03.115 "dma_device_type": 1 00:09:03.115 }, 00:09:03.115 { 00:09:03.115 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.115 "dma_device_type": 2 00:09:03.115 } 00:09:03.115 ], 00:09:03.115 "driver_specific": {} 00:09:03.115 } 00:09:03.115 ] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.115 BaseBdev3 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.115 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.375 [ 00:09:03.375 { 00:09:03.375 "name": "BaseBdev3", 00:09:03.375 "aliases": [ 00:09:03.375 "3055fff5-f7e0-4806-a5d3-9a055b652e79" 00:09:03.375 ], 00:09:03.375 "product_name": "Malloc disk", 00:09:03.375 "block_size": 512, 00:09:03.375 "num_blocks": 65536, 00:09:03.375 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:03.375 "assigned_rate_limits": { 00:09:03.375 "rw_ios_per_sec": 0, 00:09:03.375 "rw_mbytes_per_sec": 0, 00:09:03.375 "r_mbytes_per_sec": 0, 00:09:03.375 "w_mbytes_per_sec": 0 00:09:03.375 }, 00:09:03.375 "claimed": false, 00:09:03.375 "zoned": false, 00:09:03.375 "supported_io_types": { 00:09:03.375 "read": true, 00:09:03.375 "write": true, 00:09:03.375 "unmap": true, 00:09:03.375 "flush": true, 00:09:03.375 "reset": true, 00:09:03.375 "nvme_admin": false, 00:09:03.375 "nvme_io": false, 00:09:03.375 "nvme_io_md": false, 00:09:03.375 "write_zeroes": true, 00:09:03.375 "zcopy": true, 00:09:03.375 "get_zone_info": false, 00:09:03.375 "zone_management": false, 00:09:03.375 "zone_append": false, 00:09:03.375 "compare": false, 00:09:03.375 "compare_and_write": false, 00:09:03.375 "abort": true, 00:09:03.375 "seek_hole": false, 00:09:03.375 "seek_data": false, 00:09:03.375 "copy": true, 00:09:03.375 "nvme_iov_md": false 00:09:03.375 }, 00:09:03.375 "memory_domains": [ 00:09:03.375 { 00:09:03.375 "dma_device_id": "system", 00:09:03.375 "dma_device_type": 1 00:09:03.375 }, 00:09:03.375 { 00:09:03.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.375 "dma_device_type": 2 00:09:03.375 } 00:09:03.375 ], 00:09:03.375 "driver_specific": {} 00:09:03.375 } 00:09:03.375 ] 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.375 [2024-09-30 14:08:07.801044] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:03.375 [2024-09-30 14:08:07.801095] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:03.375 [2024-09-30 14:08:07.801115] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:03.375 [2024-09-30 14:08:07.802897] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.375 "name": "Existed_Raid", 00:09:03.375 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:03.375 "strip_size_kb": 0, 00:09:03.375 "state": "configuring", 00:09:03.375 "raid_level": "raid1", 00:09:03.375 "superblock": true, 00:09:03.375 "num_base_bdevs": 3, 00:09:03.375 "num_base_bdevs_discovered": 2, 00:09:03.375 "num_base_bdevs_operational": 3, 00:09:03.375 "base_bdevs_list": [ 00:09:03.375 { 00:09:03.375 "name": "BaseBdev1", 00:09:03.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.375 "is_configured": false, 00:09:03.375 "data_offset": 0, 00:09:03.375 "data_size": 0 00:09:03.375 }, 00:09:03.375 { 00:09:03.375 "name": "BaseBdev2", 00:09:03.375 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:03.375 "is_configured": true, 00:09:03.375 "data_offset": 2048, 00:09:03.375 "data_size": 63488 00:09:03.375 }, 00:09:03.375 { 00:09:03.375 "name": "BaseBdev3", 00:09:03.375 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:03.375 "is_configured": true, 00:09:03.375 "data_offset": 2048, 00:09:03.375 "data_size": 63488 00:09:03.375 } 00:09:03.375 ] 00:09:03.375 }' 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.375 14:08:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.635 [2024-09-30 14:08:08.252256] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.635 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.894 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.894 "name": "Existed_Raid", 00:09:03.894 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:03.894 "strip_size_kb": 0, 00:09:03.894 "state": "configuring", 00:09:03.894 "raid_level": "raid1", 00:09:03.894 "superblock": true, 00:09:03.894 "num_base_bdevs": 3, 00:09:03.894 "num_base_bdevs_discovered": 1, 00:09:03.894 "num_base_bdevs_operational": 3, 00:09:03.894 "base_bdevs_list": [ 00:09:03.894 { 00:09:03.894 "name": "BaseBdev1", 00:09:03.894 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.894 "is_configured": false, 00:09:03.894 "data_offset": 0, 00:09:03.894 "data_size": 0 00:09:03.894 }, 00:09:03.894 { 00:09:03.894 "name": null, 00:09:03.894 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:03.894 "is_configured": false, 00:09:03.894 "data_offset": 0, 00:09:03.894 "data_size": 63488 00:09:03.894 }, 00:09:03.894 { 00:09:03.894 "name": "BaseBdev3", 00:09:03.894 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:03.894 "is_configured": true, 00:09:03.894 "data_offset": 2048, 00:09:03.894 "data_size": 63488 00:09:03.894 } 00:09:03.894 ] 00:09:03.894 }' 00:09:03.894 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.894 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.154 [2024-09-30 14:08:08.738111] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:04.154 BaseBdev1 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.154 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.154 [ 00:09:04.154 { 00:09:04.154 "name": "BaseBdev1", 00:09:04.154 "aliases": [ 00:09:04.154 "5e87c52e-c709-4209-a132-2b356a18d61b" 00:09:04.154 ], 00:09:04.154 "product_name": "Malloc disk", 00:09:04.154 "block_size": 512, 00:09:04.154 "num_blocks": 65536, 00:09:04.154 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:04.154 "assigned_rate_limits": { 00:09:04.154 "rw_ios_per_sec": 0, 00:09:04.154 "rw_mbytes_per_sec": 0, 00:09:04.154 "r_mbytes_per_sec": 0, 00:09:04.154 "w_mbytes_per_sec": 0 00:09:04.154 }, 00:09:04.154 "claimed": true, 00:09:04.154 "claim_type": "exclusive_write", 00:09:04.154 "zoned": false, 00:09:04.154 "supported_io_types": { 00:09:04.154 "read": true, 00:09:04.154 "write": true, 00:09:04.154 "unmap": true, 00:09:04.154 "flush": true, 00:09:04.154 "reset": true, 00:09:04.154 "nvme_admin": false, 00:09:04.154 "nvme_io": false, 00:09:04.154 "nvme_io_md": false, 00:09:04.154 "write_zeroes": true, 00:09:04.154 "zcopy": true, 00:09:04.154 "get_zone_info": false, 00:09:04.154 "zone_management": false, 00:09:04.154 "zone_append": false, 00:09:04.155 "compare": false, 00:09:04.155 "compare_and_write": false, 00:09:04.155 "abort": true, 00:09:04.155 "seek_hole": false, 00:09:04.155 "seek_data": false, 00:09:04.155 "copy": true, 00:09:04.155 "nvme_iov_md": false 00:09:04.155 }, 00:09:04.155 "memory_domains": [ 00:09:04.155 { 00:09:04.155 "dma_device_id": "system", 00:09:04.155 "dma_device_type": 1 00:09:04.155 }, 00:09:04.155 { 00:09:04.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.155 "dma_device_type": 2 00:09:04.155 } 00:09:04.155 ], 00:09:04.155 "driver_specific": {} 00:09:04.155 } 00:09:04.155 ] 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.155 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.413 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.413 "name": "Existed_Raid", 00:09:04.413 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:04.413 "strip_size_kb": 0, 00:09:04.413 "state": "configuring", 00:09:04.413 "raid_level": "raid1", 00:09:04.413 "superblock": true, 00:09:04.413 "num_base_bdevs": 3, 00:09:04.413 "num_base_bdevs_discovered": 2, 00:09:04.413 "num_base_bdevs_operational": 3, 00:09:04.413 "base_bdevs_list": [ 00:09:04.413 { 00:09:04.413 "name": "BaseBdev1", 00:09:04.413 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:04.413 "is_configured": true, 00:09:04.413 "data_offset": 2048, 00:09:04.413 "data_size": 63488 00:09:04.413 }, 00:09:04.413 { 00:09:04.413 "name": null, 00:09:04.413 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:04.413 "is_configured": false, 00:09:04.413 "data_offset": 0, 00:09:04.413 "data_size": 63488 00:09:04.413 }, 00:09:04.413 { 00:09:04.413 "name": "BaseBdev3", 00:09:04.413 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:04.413 "is_configured": true, 00:09:04.413 "data_offset": 2048, 00:09:04.413 "data_size": 63488 00:09:04.413 } 00:09:04.413 ] 00:09:04.413 }' 00:09:04.413 14:08:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.413 14:08:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.672 [2024-09-30 14:08:09.281228] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.672 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.673 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.932 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.932 "name": "Existed_Raid", 00:09:04.932 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:04.932 "strip_size_kb": 0, 00:09:04.932 "state": "configuring", 00:09:04.932 "raid_level": "raid1", 00:09:04.932 "superblock": true, 00:09:04.932 "num_base_bdevs": 3, 00:09:04.932 "num_base_bdevs_discovered": 1, 00:09:04.932 "num_base_bdevs_operational": 3, 00:09:04.932 "base_bdevs_list": [ 00:09:04.932 { 00:09:04.932 "name": "BaseBdev1", 00:09:04.932 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:04.932 "is_configured": true, 00:09:04.932 "data_offset": 2048, 00:09:04.932 "data_size": 63488 00:09:04.932 }, 00:09:04.932 { 00:09:04.932 "name": null, 00:09:04.932 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:04.932 "is_configured": false, 00:09:04.932 "data_offset": 0, 00:09:04.932 "data_size": 63488 00:09:04.932 }, 00:09:04.932 { 00:09:04.932 "name": null, 00:09:04.932 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:04.932 "is_configured": false, 00:09:04.932 "data_offset": 0, 00:09:04.932 "data_size": 63488 00:09:04.932 } 00:09:04.932 ] 00:09:04.932 }' 00:09:04.932 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.932 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.192 [2024-09-30 14:08:09.736449] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.192 "name": "Existed_Raid", 00:09:05.192 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:05.192 "strip_size_kb": 0, 00:09:05.192 "state": "configuring", 00:09:05.192 "raid_level": "raid1", 00:09:05.192 "superblock": true, 00:09:05.192 "num_base_bdevs": 3, 00:09:05.192 "num_base_bdevs_discovered": 2, 00:09:05.192 "num_base_bdevs_operational": 3, 00:09:05.192 "base_bdevs_list": [ 00:09:05.192 { 00:09:05.192 "name": "BaseBdev1", 00:09:05.192 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:05.192 "is_configured": true, 00:09:05.192 "data_offset": 2048, 00:09:05.192 "data_size": 63488 00:09:05.192 }, 00:09:05.192 { 00:09:05.192 "name": null, 00:09:05.192 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:05.192 "is_configured": false, 00:09:05.192 "data_offset": 0, 00:09:05.192 "data_size": 63488 00:09:05.192 }, 00:09:05.192 { 00:09:05.192 "name": "BaseBdev3", 00:09:05.192 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:05.192 "is_configured": true, 00:09:05.192 "data_offset": 2048, 00:09:05.192 "data_size": 63488 00:09:05.192 } 00:09:05.192 ] 00:09:05.192 }' 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.192 14:08:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.761 [2024-09-30 14:08:10.191690] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.761 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.762 "name": "Existed_Raid", 00:09:05.762 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:05.762 "strip_size_kb": 0, 00:09:05.762 "state": "configuring", 00:09:05.762 "raid_level": "raid1", 00:09:05.762 "superblock": true, 00:09:05.762 "num_base_bdevs": 3, 00:09:05.762 "num_base_bdevs_discovered": 1, 00:09:05.762 "num_base_bdevs_operational": 3, 00:09:05.762 "base_bdevs_list": [ 00:09:05.762 { 00:09:05.762 "name": null, 00:09:05.762 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:05.762 "is_configured": false, 00:09:05.762 "data_offset": 0, 00:09:05.762 "data_size": 63488 00:09:05.762 }, 00:09:05.762 { 00:09:05.762 "name": null, 00:09:05.762 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:05.762 "is_configured": false, 00:09:05.762 "data_offset": 0, 00:09:05.762 "data_size": 63488 00:09:05.762 }, 00:09:05.762 { 00:09:05.762 "name": "BaseBdev3", 00:09:05.762 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:05.762 "is_configured": true, 00:09:05.762 "data_offset": 2048, 00:09:05.762 "data_size": 63488 00:09:05.762 } 00:09:05.762 ] 00:09:05.762 }' 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.762 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.022 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:06.022 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.022 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.022 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.282 [2024-09-30 14:08:10.717101] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.282 "name": "Existed_Raid", 00:09:06.282 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:06.282 "strip_size_kb": 0, 00:09:06.282 "state": "configuring", 00:09:06.282 "raid_level": "raid1", 00:09:06.282 "superblock": true, 00:09:06.282 "num_base_bdevs": 3, 00:09:06.282 "num_base_bdevs_discovered": 2, 00:09:06.282 "num_base_bdevs_operational": 3, 00:09:06.282 "base_bdevs_list": [ 00:09:06.282 { 00:09:06.282 "name": null, 00:09:06.282 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:06.282 "is_configured": false, 00:09:06.282 "data_offset": 0, 00:09:06.282 "data_size": 63488 00:09:06.282 }, 00:09:06.282 { 00:09:06.282 "name": "BaseBdev2", 00:09:06.282 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:06.282 "is_configured": true, 00:09:06.282 "data_offset": 2048, 00:09:06.282 "data_size": 63488 00:09:06.282 }, 00:09:06.282 { 00:09:06.282 "name": "BaseBdev3", 00:09:06.282 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:06.282 "is_configured": true, 00:09:06.282 "data_offset": 2048, 00:09:06.282 "data_size": 63488 00:09:06.282 } 00:09:06.282 ] 00:09:06.282 }' 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.282 14:08:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.541 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5e87c52e-c709-4209-a132-2b356a18d61b 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.800 [2024-09-30 14:08:11.262851] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:06.800 [2024-09-30 14:08:11.263006] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:06.800 [2024-09-30 14:08:11.263021] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:06.800 [2024-09-30 14:08:11.263260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:09:06.800 [2024-09-30 14:08:11.263393] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:06.800 [2024-09-30 14:08:11.263408] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:06.800 [2024-09-30 14:08:11.263522] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:06.800 NewBaseBdev 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.800 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.800 [ 00:09:06.800 { 00:09:06.800 "name": "NewBaseBdev", 00:09:06.800 "aliases": [ 00:09:06.800 "5e87c52e-c709-4209-a132-2b356a18d61b" 00:09:06.800 ], 00:09:06.800 "product_name": "Malloc disk", 00:09:06.800 "block_size": 512, 00:09:06.800 "num_blocks": 65536, 00:09:06.800 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:06.800 "assigned_rate_limits": { 00:09:06.800 "rw_ios_per_sec": 0, 00:09:06.800 "rw_mbytes_per_sec": 0, 00:09:06.800 "r_mbytes_per_sec": 0, 00:09:06.800 "w_mbytes_per_sec": 0 00:09:06.800 }, 00:09:06.800 "claimed": true, 00:09:06.800 "claim_type": "exclusive_write", 00:09:06.800 "zoned": false, 00:09:06.800 "supported_io_types": { 00:09:06.800 "read": true, 00:09:06.800 "write": true, 00:09:06.800 "unmap": true, 00:09:06.800 "flush": true, 00:09:06.800 "reset": true, 00:09:06.800 "nvme_admin": false, 00:09:06.800 "nvme_io": false, 00:09:06.800 "nvme_io_md": false, 00:09:06.800 "write_zeroes": true, 00:09:06.800 "zcopy": true, 00:09:06.800 "get_zone_info": false, 00:09:06.800 "zone_management": false, 00:09:06.800 "zone_append": false, 00:09:06.800 "compare": false, 00:09:06.800 "compare_and_write": false, 00:09:06.800 "abort": true, 00:09:06.800 "seek_hole": false, 00:09:06.800 "seek_data": false, 00:09:06.800 "copy": true, 00:09:06.800 "nvme_iov_md": false 00:09:06.800 }, 00:09:06.800 "memory_domains": [ 00:09:06.800 { 00:09:06.800 "dma_device_id": "system", 00:09:06.800 "dma_device_type": 1 00:09:06.800 }, 00:09:06.800 { 00:09:06.800 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.800 "dma_device_type": 2 00:09:06.800 } 00:09:06.800 ], 00:09:06.800 "driver_specific": {} 00:09:06.801 } 00:09:06.801 ] 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.801 "name": "Existed_Raid", 00:09:06.801 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:06.801 "strip_size_kb": 0, 00:09:06.801 "state": "online", 00:09:06.801 "raid_level": "raid1", 00:09:06.801 "superblock": true, 00:09:06.801 "num_base_bdevs": 3, 00:09:06.801 "num_base_bdevs_discovered": 3, 00:09:06.801 "num_base_bdevs_operational": 3, 00:09:06.801 "base_bdevs_list": [ 00:09:06.801 { 00:09:06.801 "name": "NewBaseBdev", 00:09:06.801 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:06.801 "is_configured": true, 00:09:06.801 "data_offset": 2048, 00:09:06.801 "data_size": 63488 00:09:06.801 }, 00:09:06.801 { 00:09:06.801 "name": "BaseBdev2", 00:09:06.801 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:06.801 "is_configured": true, 00:09:06.801 "data_offset": 2048, 00:09:06.801 "data_size": 63488 00:09:06.801 }, 00:09:06.801 { 00:09:06.801 "name": "BaseBdev3", 00:09:06.801 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:06.801 "is_configured": true, 00:09:06.801 "data_offset": 2048, 00:09:06.801 "data_size": 63488 00:09:06.801 } 00:09:06.801 ] 00:09:06.801 }' 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.801 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:07.369 [2024-09-30 14:08:11.758353] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.369 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:07.369 "name": "Existed_Raid", 00:09:07.369 "aliases": [ 00:09:07.369 "328b861a-290b-4cef-be37-16210984c952" 00:09:07.369 ], 00:09:07.369 "product_name": "Raid Volume", 00:09:07.369 "block_size": 512, 00:09:07.369 "num_blocks": 63488, 00:09:07.369 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:07.369 "assigned_rate_limits": { 00:09:07.369 "rw_ios_per_sec": 0, 00:09:07.369 "rw_mbytes_per_sec": 0, 00:09:07.369 "r_mbytes_per_sec": 0, 00:09:07.369 "w_mbytes_per_sec": 0 00:09:07.369 }, 00:09:07.369 "claimed": false, 00:09:07.369 "zoned": false, 00:09:07.369 "supported_io_types": { 00:09:07.369 "read": true, 00:09:07.369 "write": true, 00:09:07.369 "unmap": false, 00:09:07.369 "flush": false, 00:09:07.369 "reset": true, 00:09:07.369 "nvme_admin": false, 00:09:07.369 "nvme_io": false, 00:09:07.369 "nvme_io_md": false, 00:09:07.369 "write_zeroes": true, 00:09:07.369 "zcopy": false, 00:09:07.369 "get_zone_info": false, 00:09:07.369 "zone_management": false, 00:09:07.369 "zone_append": false, 00:09:07.369 "compare": false, 00:09:07.369 "compare_and_write": false, 00:09:07.369 "abort": false, 00:09:07.369 "seek_hole": false, 00:09:07.369 "seek_data": false, 00:09:07.369 "copy": false, 00:09:07.369 "nvme_iov_md": false 00:09:07.369 }, 00:09:07.369 "memory_domains": [ 00:09:07.369 { 00:09:07.369 "dma_device_id": "system", 00:09:07.370 "dma_device_type": 1 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.370 "dma_device_type": 2 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "dma_device_id": "system", 00:09:07.370 "dma_device_type": 1 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.370 "dma_device_type": 2 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "dma_device_id": "system", 00:09:07.370 "dma_device_type": 1 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.370 "dma_device_type": 2 00:09:07.370 } 00:09:07.370 ], 00:09:07.370 "driver_specific": { 00:09:07.370 "raid": { 00:09:07.370 "uuid": "328b861a-290b-4cef-be37-16210984c952", 00:09:07.370 "strip_size_kb": 0, 00:09:07.370 "state": "online", 00:09:07.370 "raid_level": "raid1", 00:09:07.370 "superblock": true, 00:09:07.370 "num_base_bdevs": 3, 00:09:07.370 "num_base_bdevs_discovered": 3, 00:09:07.370 "num_base_bdevs_operational": 3, 00:09:07.370 "base_bdevs_list": [ 00:09:07.370 { 00:09:07.370 "name": "NewBaseBdev", 00:09:07.370 "uuid": "5e87c52e-c709-4209-a132-2b356a18d61b", 00:09:07.370 "is_configured": true, 00:09:07.370 "data_offset": 2048, 00:09:07.370 "data_size": 63488 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "name": "BaseBdev2", 00:09:07.370 "uuid": "37c7cc68-b2f4-4312-ac2a-d17b9b4965f1", 00:09:07.370 "is_configured": true, 00:09:07.370 "data_offset": 2048, 00:09:07.370 "data_size": 63488 00:09:07.370 }, 00:09:07.370 { 00:09:07.370 "name": "BaseBdev3", 00:09:07.370 "uuid": "3055fff5-f7e0-4806-a5d3-9a055b652e79", 00:09:07.370 "is_configured": true, 00:09:07.370 "data_offset": 2048, 00:09:07.370 "data_size": 63488 00:09:07.370 } 00:09:07.370 ] 00:09:07.370 } 00:09:07.370 } 00:09:07.370 }' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:07.370 BaseBdev2 00:09:07.370 BaseBdev3' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.370 14:08:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.370 [2024-09-30 14:08:12.013599] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:07.370 [2024-09-30 14:08:12.013624] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:07.370 [2024-09-30 14:08:12.013689] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:07.370 [2024-09-30 14:08:12.013916] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:07.370 [2024-09-30 14:08:12.013936] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80094 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 80094 ']' 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 80094 00:09:07.370 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80094 00:09:07.629 killing process with pid 80094 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80094' 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 80094 00:09:07.629 [2024-09-30 14:08:12.068640] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:07.629 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 80094 00:09:07.629 [2024-09-30 14:08:12.098600] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:07.888 14:08:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:07.888 00:09:07.888 real 0m8.781s 00:09:07.888 user 0m14.939s 00:09:07.888 sys 0m1.830s 00:09:07.888 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:07.888 14:08:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.888 ************************************ 00:09:07.888 END TEST raid_state_function_test_sb 00:09:07.888 ************************************ 00:09:07.888 14:08:12 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:07.888 14:08:12 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:07.888 14:08:12 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:07.888 14:08:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:07.888 ************************************ 00:09:07.888 START TEST raid_superblock_test 00:09:07.888 ************************************ 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 3 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=80698 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 80698 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 80698 ']' 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:07.888 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:07.888 14:08:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.888 [2024-09-30 14:08:12.503275] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:07.888 [2024-09-30 14:08:12.503404] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80698 ] 00:09:08.147 [2024-09-30 14:08:12.633276] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:08.147 [2024-09-30 14:08:12.662753] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:08.147 [2024-09-30 14:08:12.707036] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:08.147 [2024-09-30 14:08:12.747956] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:08.147 [2024-09-30 14:08:12.747992] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.716 malloc1 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.716 [2024-09-30 14:08:13.361025] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:08.716 [2024-09-30 14:08:13.361098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.716 [2024-09-30 14:08:13.361116] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:08.716 [2024-09-30 14:08:13.361127] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.716 [2024-09-30 14:08:13.363088] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.716 [2024-09-30 14:08:13.363123] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:08.716 pt1 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.716 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 malloc2 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 [2024-09-30 14:08:13.405144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:08.975 [2024-09-30 14:08:13.405273] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.975 [2024-09-30 14:08:13.405322] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:08.975 [2024-09-30 14:08:13.405349] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.975 [2024-09-30 14:08:13.410020] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.975 [2024-09-30 14:08:13.410080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:08.975 pt2 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 malloc3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 [2024-09-30 14:08:13.435908] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:08.975 [2024-09-30 14:08:13.435955] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.975 [2024-09-30 14:08:13.435975] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:08.975 [2024-09-30 14:08:13.435983] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.975 [2024-09-30 14:08:13.438040] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.975 [2024-09-30 14:08:13.438072] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:08.975 pt3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 [2024-09-30 14:08:13.447940] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:08.975 [2024-09-30 14:08:13.449626] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:08.975 [2024-09-30 14:08:13.449684] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:08.975 [2024-09-30 14:08:13.449811] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:09:08.975 [2024-09-30 14:08:13.449824] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:08.975 [2024-09-30 14:08:13.450057] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:08.975 [2024-09-30 14:08:13.450184] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:09:08.975 [2024-09-30 14:08:13.450198] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:09:08.975 [2024-09-30 14:08:13.450307] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.975 "name": "raid_bdev1", 00:09:08.975 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:08.975 "strip_size_kb": 0, 00:09:08.975 "state": "online", 00:09:08.975 "raid_level": "raid1", 00:09:08.975 "superblock": true, 00:09:08.975 "num_base_bdevs": 3, 00:09:08.975 "num_base_bdevs_discovered": 3, 00:09:08.975 "num_base_bdevs_operational": 3, 00:09:08.975 "base_bdevs_list": [ 00:09:08.975 { 00:09:08.975 "name": "pt1", 00:09:08.975 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:08.975 "is_configured": true, 00:09:08.975 "data_offset": 2048, 00:09:08.975 "data_size": 63488 00:09:08.975 }, 00:09:08.975 { 00:09:08.975 "name": "pt2", 00:09:08.975 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:08.975 "is_configured": true, 00:09:08.975 "data_offset": 2048, 00:09:08.975 "data_size": 63488 00:09:08.975 }, 00:09:08.975 { 00:09:08.975 "name": "pt3", 00:09:08.975 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:08.975 "is_configured": true, 00:09:08.975 "data_offset": 2048, 00:09:08.975 "data_size": 63488 00:09:08.975 } 00:09:08.975 ] 00:09:08.975 }' 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.975 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.234 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.234 [2024-09-30 14:08:13.883441] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:09.494 "name": "raid_bdev1", 00:09:09.494 "aliases": [ 00:09:09.494 "5a06a384-b99f-4adc-b8de-9313e6ea9bd3" 00:09:09.494 ], 00:09:09.494 "product_name": "Raid Volume", 00:09:09.494 "block_size": 512, 00:09:09.494 "num_blocks": 63488, 00:09:09.494 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:09.494 "assigned_rate_limits": { 00:09:09.494 "rw_ios_per_sec": 0, 00:09:09.494 "rw_mbytes_per_sec": 0, 00:09:09.494 "r_mbytes_per_sec": 0, 00:09:09.494 "w_mbytes_per_sec": 0 00:09:09.494 }, 00:09:09.494 "claimed": false, 00:09:09.494 "zoned": false, 00:09:09.494 "supported_io_types": { 00:09:09.494 "read": true, 00:09:09.494 "write": true, 00:09:09.494 "unmap": false, 00:09:09.494 "flush": false, 00:09:09.494 "reset": true, 00:09:09.494 "nvme_admin": false, 00:09:09.494 "nvme_io": false, 00:09:09.494 "nvme_io_md": false, 00:09:09.494 "write_zeroes": true, 00:09:09.494 "zcopy": false, 00:09:09.494 "get_zone_info": false, 00:09:09.494 "zone_management": false, 00:09:09.494 "zone_append": false, 00:09:09.494 "compare": false, 00:09:09.494 "compare_and_write": false, 00:09:09.494 "abort": false, 00:09:09.494 "seek_hole": false, 00:09:09.494 "seek_data": false, 00:09:09.494 "copy": false, 00:09:09.494 "nvme_iov_md": false 00:09:09.494 }, 00:09:09.494 "memory_domains": [ 00:09:09.494 { 00:09:09.494 "dma_device_id": "system", 00:09:09.494 "dma_device_type": 1 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.494 "dma_device_type": 2 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "dma_device_id": "system", 00:09:09.494 "dma_device_type": 1 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.494 "dma_device_type": 2 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "dma_device_id": "system", 00:09:09.494 "dma_device_type": 1 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.494 "dma_device_type": 2 00:09:09.494 } 00:09:09.494 ], 00:09:09.494 "driver_specific": { 00:09:09.494 "raid": { 00:09:09.494 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:09.494 "strip_size_kb": 0, 00:09:09.494 "state": "online", 00:09:09.494 "raid_level": "raid1", 00:09:09.494 "superblock": true, 00:09:09.494 "num_base_bdevs": 3, 00:09:09.494 "num_base_bdevs_discovered": 3, 00:09:09.494 "num_base_bdevs_operational": 3, 00:09:09.494 "base_bdevs_list": [ 00:09:09.494 { 00:09:09.494 "name": "pt1", 00:09:09.494 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:09.494 "is_configured": true, 00:09:09.494 "data_offset": 2048, 00:09:09.494 "data_size": 63488 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "name": "pt2", 00:09:09.494 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:09.494 "is_configured": true, 00:09:09.494 "data_offset": 2048, 00:09:09.494 "data_size": 63488 00:09:09.494 }, 00:09:09.494 { 00:09:09.494 "name": "pt3", 00:09:09.494 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:09.494 "is_configured": true, 00:09:09.494 "data_offset": 2048, 00:09:09.494 "data_size": 63488 00:09:09.494 } 00:09:09.494 ] 00:09:09.494 } 00:09:09.494 } 00:09:09.494 }' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:09.494 pt2 00:09:09.494 pt3' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.494 14:08:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.494 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:09.495 [2024-09-30 14:08:14.111043] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=5a06a384-b99f-4adc-b8de-9313e6ea9bd3 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 5a06a384-b99f-4adc-b8de-9313e6ea9bd3 ']' 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.495 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 [2024-09-30 14:08:14.150812] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:09.755 [2024-09-30 14:08:14.150848] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:09.755 [2024-09-30 14:08:14.150922] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:09.755 [2024-09-30 14:08:14.151004] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:09.755 [2024-09-30 14:08:14.151016] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 [2024-09-30 14:08:14.298581] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:09.755 [2024-09-30 14:08:14.300381] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:09.755 [2024-09-30 14:08:14.300433] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:09.755 [2024-09-30 14:08:14.300493] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:09.755 [2024-09-30 14:08:14.300548] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:09.755 [2024-09-30 14:08:14.300569] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:09.755 [2024-09-30 14:08:14.300582] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:09.755 [2024-09-30 14:08:14.300592] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:09:09.755 request: 00:09:09.755 { 00:09:09.755 "name": "raid_bdev1", 00:09:09.755 "raid_level": "raid1", 00:09:09.755 "base_bdevs": [ 00:09:09.755 "malloc1", 00:09:09.755 "malloc2", 00:09:09.755 "malloc3" 00:09:09.755 ], 00:09:09.755 "superblock": false, 00:09:09.755 "method": "bdev_raid_create", 00:09:09.755 "req_id": 1 00:09:09.755 } 00:09:09.755 Got JSON-RPC error response 00:09:09.755 response: 00:09:09.755 { 00:09:09.755 "code": -17, 00:09:09.755 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:09.755 } 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.755 [2024-09-30 14:08:14.350444] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:09.755 [2024-09-30 14:08:14.350507] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.755 [2024-09-30 14:08:14.350527] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:09.755 [2024-09-30 14:08:14.350536] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.755 [2024-09-30 14:08:14.352626] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.755 [2024-09-30 14:08:14.352659] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:09.755 [2024-09-30 14:08:14.352725] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:09.755 [2024-09-30 14:08:14.352770] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:09.755 pt1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.755 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.756 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.756 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.015 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.015 "name": "raid_bdev1", 00:09:10.015 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:10.015 "strip_size_kb": 0, 00:09:10.015 "state": "configuring", 00:09:10.015 "raid_level": "raid1", 00:09:10.015 "superblock": true, 00:09:10.015 "num_base_bdevs": 3, 00:09:10.015 "num_base_bdevs_discovered": 1, 00:09:10.015 "num_base_bdevs_operational": 3, 00:09:10.015 "base_bdevs_list": [ 00:09:10.015 { 00:09:10.015 "name": "pt1", 00:09:10.015 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:10.015 "is_configured": true, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "name": null, 00:09:10.015 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:10.015 "is_configured": false, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "name": null, 00:09:10.015 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:10.015 "is_configured": false, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 } 00:09:10.015 ] 00:09:10.015 }' 00:09:10.015 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.015 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.275 [2024-09-30 14:08:14.821652] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:10.275 [2024-09-30 14:08:14.821722] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:10.275 [2024-09-30 14:08:14.821746] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:10.275 [2024-09-30 14:08:14.821755] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:10.275 [2024-09-30 14:08:14.822124] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:10.275 [2024-09-30 14:08:14.822151] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:10.275 [2024-09-30 14:08:14.822223] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:10.275 [2024-09-30 14:08:14.822244] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:10.275 pt2 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.275 [2024-09-30 14:08:14.833636] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.275 "name": "raid_bdev1", 00:09:10.275 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:10.275 "strip_size_kb": 0, 00:09:10.275 "state": "configuring", 00:09:10.275 "raid_level": "raid1", 00:09:10.275 "superblock": true, 00:09:10.275 "num_base_bdevs": 3, 00:09:10.275 "num_base_bdevs_discovered": 1, 00:09:10.275 "num_base_bdevs_operational": 3, 00:09:10.275 "base_bdevs_list": [ 00:09:10.275 { 00:09:10.275 "name": "pt1", 00:09:10.275 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:10.275 "is_configured": true, 00:09:10.275 "data_offset": 2048, 00:09:10.275 "data_size": 63488 00:09:10.275 }, 00:09:10.275 { 00:09:10.275 "name": null, 00:09:10.275 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:10.275 "is_configured": false, 00:09:10.275 "data_offset": 0, 00:09:10.275 "data_size": 63488 00:09:10.275 }, 00:09:10.275 { 00:09:10.275 "name": null, 00:09:10.275 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:10.275 "is_configured": false, 00:09:10.275 "data_offset": 2048, 00:09:10.275 "data_size": 63488 00:09:10.275 } 00:09:10.275 ] 00:09:10.275 }' 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.275 14:08:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.844 [2024-09-30 14:08:15.308770] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:10.844 [2024-09-30 14:08:15.308827] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:10.844 [2024-09-30 14:08:15.308843] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:10.844 [2024-09-30 14:08:15.308863] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:10.844 [2024-09-30 14:08:15.309221] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:10.844 [2024-09-30 14:08:15.309247] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:10.844 [2024-09-30 14:08:15.309311] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:10.844 [2024-09-30 14:08:15.309339] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:10.844 pt2 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.844 [2024-09-30 14:08:15.320752] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:10.844 [2024-09-30 14:08:15.320796] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:10.844 [2024-09-30 14:08:15.320808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:10.844 [2024-09-30 14:08:15.320817] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:10.844 [2024-09-30 14:08:15.321098] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:10.844 [2024-09-30 14:08:15.321128] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:10.844 [2024-09-30 14:08:15.321179] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:10.844 [2024-09-30 14:08:15.321197] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:10.844 [2024-09-30 14:08:15.321283] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:10.844 [2024-09-30 14:08:15.321295] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:10.844 [2024-09-30 14:08:15.321504] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:10.844 [2024-09-30 14:08:15.321618] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:10.844 [2024-09-30 14:08:15.321629] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:10.844 [2024-09-30 14:08:15.321721] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:10.844 pt3 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.844 "name": "raid_bdev1", 00:09:10.844 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:10.844 "strip_size_kb": 0, 00:09:10.844 "state": "online", 00:09:10.844 "raid_level": "raid1", 00:09:10.844 "superblock": true, 00:09:10.844 "num_base_bdevs": 3, 00:09:10.844 "num_base_bdevs_discovered": 3, 00:09:10.844 "num_base_bdevs_operational": 3, 00:09:10.844 "base_bdevs_list": [ 00:09:10.844 { 00:09:10.844 "name": "pt1", 00:09:10.844 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:10.844 "is_configured": true, 00:09:10.844 "data_offset": 2048, 00:09:10.844 "data_size": 63488 00:09:10.844 }, 00:09:10.844 { 00:09:10.844 "name": "pt2", 00:09:10.844 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:10.844 "is_configured": true, 00:09:10.844 "data_offset": 2048, 00:09:10.844 "data_size": 63488 00:09:10.844 }, 00:09:10.844 { 00:09:10.844 "name": "pt3", 00:09:10.844 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:10.844 "is_configured": true, 00:09:10.844 "data_offset": 2048, 00:09:10.844 "data_size": 63488 00:09:10.844 } 00:09:10.844 ] 00:09:10.844 }' 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.844 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.104 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.104 [2024-09-30 14:08:15.752266] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:11.363 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.363 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:11.363 "name": "raid_bdev1", 00:09:11.363 "aliases": [ 00:09:11.363 "5a06a384-b99f-4adc-b8de-9313e6ea9bd3" 00:09:11.363 ], 00:09:11.363 "product_name": "Raid Volume", 00:09:11.363 "block_size": 512, 00:09:11.363 "num_blocks": 63488, 00:09:11.363 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:11.363 "assigned_rate_limits": { 00:09:11.363 "rw_ios_per_sec": 0, 00:09:11.363 "rw_mbytes_per_sec": 0, 00:09:11.363 "r_mbytes_per_sec": 0, 00:09:11.363 "w_mbytes_per_sec": 0 00:09:11.363 }, 00:09:11.363 "claimed": false, 00:09:11.363 "zoned": false, 00:09:11.363 "supported_io_types": { 00:09:11.363 "read": true, 00:09:11.363 "write": true, 00:09:11.363 "unmap": false, 00:09:11.363 "flush": false, 00:09:11.363 "reset": true, 00:09:11.363 "nvme_admin": false, 00:09:11.363 "nvme_io": false, 00:09:11.363 "nvme_io_md": false, 00:09:11.363 "write_zeroes": true, 00:09:11.363 "zcopy": false, 00:09:11.363 "get_zone_info": false, 00:09:11.363 "zone_management": false, 00:09:11.363 "zone_append": false, 00:09:11.363 "compare": false, 00:09:11.363 "compare_and_write": false, 00:09:11.363 "abort": false, 00:09:11.363 "seek_hole": false, 00:09:11.363 "seek_data": false, 00:09:11.363 "copy": false, 00:09:11.363 "nvme_iov_md": false 00:09:11.363 }, 00:09:11.363 "memory_domains": [ 00:09:11.363 { 00:09:11.363 "dma_device_id": "system", 00:09:11.363 "dma_device_type": 1 00:09:11.363 }, 00:09:11.363 { 00:09:11.363 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.363 "dma_device_type": 2 00:09:11.363 }, 00:09:11.363 { 00:09:11.363 "dma_device_id": "system", 00:09:11.363 "dma_device_type": 1 00:09:11.363 }, 00:09:11.363 { 00:09:11.363 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.363 "dma_device_type": 2 00:09:11.363 }, 00:09:11.363 { 00:09:11.363 "dma_device_id": "system", 00:09:11.363 "dma_device_type": 1 00:09:11.363 }, 00:09:11.363 { 00:09:11.363 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.363 "dma_device_type": 2 00:09:11.363 } 00:09:11.363 ], 00:09:11.363 "driver_specific": { 00:09:11.363 "raid": { 00:09:11.363 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:11.363 "strip_size_kb": 0, 00:09:11.363 "state": "online", 00:09:11.364 "raid_level": "raid1", 00:09:11.364 "superblock": true, 00:09:11.364 "num_base_bdevs": 3, 00:09:11.364 "num_base_bdevs_discovered": 3, 00:09:11.364 "num_base_bdevs_operational": 3, 00:09:11.364 "base_bdevs_list": [ 00:09:11.364 { 00:09:11.364 "name": "pt1", 00:09:11.364 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:11.364 "is_configured": true, 00:09:11.364 "data_offset": 2048, 00:09:11.364 "data_size": 63488 00:09:11.364 }, 00:09:11.364 { 00:09:11.364 "name": "pt2", 00:09:11.364 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:11.364 "is_configured": true, 00:09:11.364 "data_offset": 2048, 00:09:11.364 "data_size": 63488 00:09:11.364 }, 00:09:11.364 { 00:09:11.364 "name": "pt3", 00:09:11.364 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:11.364 "is_configured": true, 00:09:11.364 "data_offset": 2048, 00:09:11.364 "data_size": 63488 00:09:11.364 } 00:09:11.364 ] 00:09:11.364 } 00:09:11.364 } 00:09:11.364 }' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:11.364 pt2 00:09:11.364 pt3' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.364 14:08:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.364 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.623 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:11.623 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.624 [2024-09-30 14:08:16.039758] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 5a06a384-b99f-4adc-b8de-9313e6ea9bd3 '!=' 5a06a384-b99f-4adc-b8de-9313e6ea9bd3 ']' 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.624 [2024-09-30 14:08:16.071512] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.624 "name": "raid_bdev1", 00:09:11.624 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:11.624 "strip_size_kb": 0, 00:09:11.624 "state": "online", 00:09:11.624 "raid_level": "raid1", 00:09:11.624 "superblock": true, 00:09:11.624 "num_base_bdevs": 3, 00:09:11.624 "num_base_bdevs_discovered": 2, 00:09:11.624 "num_base_bdevs_operational": 2, 00:09:11.624 "base_bdevs_list": [ 00:09:11.624 { 00:09:11.624 "name": null, 00:09:11.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.624 "is_configured": false, 00:09:11.624 "data_offset": 0, 00:09:11.624 "data_size": 63488 00:09:11.624 }, 00:09:11.624 { 00:09:11.624 "name": "pt2", 00:09:11.624 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:11.624 "is_configured": true, 00:09:11.624 "data_offset": 2048, 00:09:11.624 "data_size": 63488 00:09:11.624 }, 00:09:11.624 { 00:09:11.624 "name": "pt3", 00:09:11.624 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:11.624 "is_configured": true, 00:09:11.624 "data_offset": 2048, 00:09:11.624 "data_size": 63488 00:09:11.624 } 00:09:11.624 ] 00:09:11.624 }' 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.624 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.884 [2024-09-30 14:08:16.518687] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:11.884 [2024-09-30 14:08:16.518715] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:11.884 [2024-09-30 14:08:16.518779] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:11.884 [2024-09-30 14:08:16.518833] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:11.884 [2024-09-30 14:08:16.518845] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.884 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:12.143 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.144 [2024-09-30 14:08:16.586567] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:12.144 [2024-09-30 14:08:16.586616] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:12.144 [2024-09-30 14:08:16.586631] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:12.144 [2024-09-30 14:08:16.586641] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:12.144 [2024-09-30 14:08:16.588605] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:12.144 [2024-09-30 14:08:16.588641] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:12.144 [2024-09-30 14:08:16.588702] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:12.144 [2024-09-30 14:08:16.588740] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:12.144 pt2 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.144 "name": "raid_bdev1", 00:09:12.144 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:12.144 "strip_size_kb": 0, 00:09:12.144 "state": "configuring", 00:09:12.144 "raid_level": "raid1", 00:09:12.144 "superblock": true, 00:09:12.144 "num_base_bdevs": 3, 00:09:12.144 "num_base_bdevs_discovered": 1, 00:09:12.144 "num_base_bdevs_operational": 2, 00:09:12.144 "base_bdevs_list": [ 00:09:12.144 { 00:09:12.144 "name": null, 00:09:12.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.144 "is_configured": false, 00:09:12.144 "data_offset": 2048, 00:09:12.144 "data_size": 63488 00:09:12.144 }, 00:09:12.144 { 00:09:12.144 "name": "pt2", 00:09:12.144 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:12.144 "is_configured": true, 00:09:12.144 "data_offset": 2048, 00:09:12.144 "data_size": 63488 00:09:12.144 }, 00:09:12.144 { 00:09:12.144 "name": null, 00:09:12.144 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:12.144 "is_configured": false, 00:09:12.144 "data_offset": 2048, 00:09:12.144 "data_size": 63488 00:09:12.144 } 00:09:12.144 ] 00:09:12.144 }' 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.144 14:08:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.403 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.403 [2024-09-30 14:08:17.037892] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:12.403 [2024-09-30 14:08:17.037954] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:12.403 [2024-09-30 14:08:17.037973] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:09:12.404 [2024-09-30 14:08:17.037986] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:12.404 [2024-09-30 14:08:17.038334] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:12.404 [2024-09-30 14:08:17.038359] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:12.404 [2024-09-30 14:08:17.038424] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:12.404 [2024-09-30 14:08:17.038451] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:12.404 [2024-09-30 14:08:17.038541] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:12.404 [2024-09-30 14:08:17.038558] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:12.404 [2024-09-30 14:08:17.038798] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:12.404 [2024-09-30 14:08:17.038915] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:12.404 [2024-09-30 14:08:17.038927] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:12.404 [2024-09-30 14:08:17.039027] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:12.404 pt3 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.404 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.663 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.663 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.663 "name": "raid_bdev1", 00:09:12.663 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:12.663 "strip_size_kb": 0, 00:09:12.663 "state": "online", 00:09:12.663 "raid_level": "raid1", 00:09:12.663 "superblock": true, 00:09:12.663 "num_base_bdevs": 3, 00:09:12.663 "num_base_bdevs_discovered": 2, 00:09:12.663 "num_base_bdevs_operational": 2, 00:09:12.663 "base_bdevs_list": [ 00:09:12.663 { 00:09:12.663 "name": null, 00:09:12.663 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.663 "is_configured": false, 00:09:12.663 "data_offset": 2048, 00:09:12.663 "data_size": 63488 00:09:12.663 }, 00:09:12.663 { 00:09:12.663 "name": "pt2", 00:09:12.663 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:12.663 "is_configured": true, 00:09:12.663 "data_offset": 2048, 00:09:12.663 "data_size": 63488 00:09:12.663 }, 00:09:12.663 { 00:09:12.663 "name": "pt3", 00:09:12.663 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:12.663 "is_configured": true, 00:09:12.663 "data_offset": 2048, 00:09:12.663 "data_size": 63488 00:09:12.663 } 00:09:12.663 ] 00:09:12.663 }' 00:09:12.663 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.663 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.923 [2024-09-30 14:08:17.529042] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:12.923 [2024-09-30 14:08:17.529072] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:12.923 [2024-09-30 14:08:17.529135] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:12.923 [2024-09-30 14:08:17.529190] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:12.923 [2024-09-30 14:08:17.529205] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:12.923 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.182 [2024-09-30 14:08:17.604915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:13.182 [2024-09-30 14:08:17.604966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:13.182 [2024-09-30 14:08:17.604984] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:09:13.182 [2024-09-30 14:08:17.604993] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:13.182 [2024-09-30 14:08:17.607001] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:13.182 [2024-09-30 14:08:17.607035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:13.182 [2024-09-30 14:08:17.607100] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:13.182 [2024-09-30 14:08:17.607135] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:13.182 [2024-09-30 14:08:17.607224] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:13.182 [2024-09-30 14:08:17.607239] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:13.182 [2024-09-30 14:08:17.607254] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:09:13.182 [2024-09-30 14:08:17.607280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:13.182 pt1 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.182 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.183 "name": "raid_bdev1", 00:09:13.183 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:13.183 "strip_size_kb": 0, 00:09:13.183 "state": "configuring", 00:09:13.183 "raid_level": "raid1", 00:09:13.183 "superblock": true, 00:09:13.183 "num_base_bdevs": 3, 00:09:13.183 "num_base_bdevs_discovered": 1, 00:09:13.183 "num_base_bdevs_operational": 2, 00:09:13.183 "base_bdevs_list": [ 00:09:13.183 { 00:09:13.183 "name": null, 00:09:13.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:13.183 "is_configured": false, 00:09:13.183 "data_offset": 2048, 00:09:13.183 "data_size": 63488 00:09:13.183 }, 00:09:13.183 { 00:09:13.183 "name": "pt2", 00:09:13.183 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:13.183 "is_configured": true, 00:09:13.183 "data_offset": 2048, 00:09:13.183 "data_size": 63488 00:09:13.183 }, 00:09:13.183 { 00:09:13.183 "name": null, 00:09:13.183 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:13.183 "is_configured": false, 00:09:13.183 "data_offset": 2048, 00:09:13.183 "data_size": 63488 00:09:13.183 } 00:09:13.183 ] 00:09:13.183 }' 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.183 14:08:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.442 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:13.442 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:13.442 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.442 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.442 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.702 [2024-09-30 14:08:18.124060] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:13.702 [2024-09-30 14:08:18.124116] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:13.702 [2024-09-30 14:08:18.124135] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:09:13.702 [2024-09-30 14:08:18.124143] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:13.702 [2024-09-30 14:08:18.124498] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:13.702 [2024-09-30 14:08:18.124526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:13.702 [2024-09-30 14:08:18.124598] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:13.702 [2024-09-30 14:08:18.124640] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:13.702 [2024-09-30 14:08:18.124730] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:09:13.702 [2024-09-30 14:08:18.124744] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:13.702 [2024-09-30 14:08:18.124946] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:09:13.702 [2024-09-30 14:08:18.125072] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:09:13.702 [2024-09-30 14:08:18.125088] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:09:13.702 [2024-09-30 14:08:18.125184] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:13.702 pt3 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:13.702 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.703 "name": "raid_bdev1", 00:09:13.703 "uuid": "5a06a384-b99f-4adc-b8de-9313e6ea9bd3", 00:09:13.703 "strip_size_kb": 0, 00:09:13.703 "state": "online", 00:09:13.703 "raid_level": "raid1", 00:09:13.703 "superblock": true, 00:09:13.703 "num_base_bdevs": 3, 00:09:13.703 "num_base_bdevs_discovered": 2, 00:09:13.703 "num_base_bdevs_operational": 2, 00:09:13.703 "base_bdevs_list": [ 00:09:13.703 { 00:09:13.703 "name": null, 00:09:13.703 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:13.703 "is_configured": false, 00:09:13.703 "data_offset": 2048, 00:09:13.703 "data_size": 63488 00:09:13.703 }, 00:09:13.703 { 00:09:13.703 "name": "pt2", 00:09:13.703 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:13.703 "is_configured": true, 00:09:13.703 "data_offset": 2048, 00:09:13.703 "data_size": 63488 00:09:13.703 }, 00:09:13.703 { 00:09:13.703 "name": "pt3", 00:09:13.703 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:13.703 "is_configured": true, 00:09:13.703 "data_offset": 2048, 00:09:13.703 "data_size": 63488 00:09:13.703 } 00:09:13.703 ] 00:09:13.703 }' 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.703 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.963 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:13.963 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.963 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.963 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:13.963 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:14.223 [2024-09-30 14:08:18.631391] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 5a06a384-b99f-4adc-b8de-9313e6ea9bd3 '!=' 5a06a384-b99f-4adc-b8de-9313e6ea9bd3 ']' 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 80698 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 80698 ']' 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 80698 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80698 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80698' 00:09:14.223 killing process with pid 80698 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 80698 00:09:14.223 [2024-09-30 14:08:18.719188] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:14.223 [2024-09-30 14:08:18.719291] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:14.223 [2024-09-30 14:08:18.719348] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:14.223 [2024-09-30 14:08:18.719360] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:09:14.223 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 80698 00:09:14.223 [2024-09-30 14:08:18.752473] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:14.483 14:08:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:14.483 00:09:14.483 real 0m6.576s 00:09:14.483 user 0m11.016s 00:09:14.483 sys 0m1.375s 00:09:14.483 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:14.483 14:08:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.483 ************************************ 00:09:14.483 END TEST raid_superblock_test 00:09:14.483 ************************************ 00:09:14.483 14:08:19 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:14.483 14:08:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:14.483 14:08:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:14.483 14:08:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:14.483 ************************************ 00:09:14.483 START TEST raid_read_error_test 00:09:14.483 ************************************ 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 read 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4NgMFUbOgf 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81127 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81127 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 81127 ']' 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:14.483 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:14.483 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.743 [2024-09-30 14:08:19.174387] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:14.743 [2024-09-30 14:08:19.174605] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81127 ] 00:09:14.743 [2024-09-30 14:08:19.304998] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:14.743 [2024-09-30 14:08:19.320288] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:14.743 [2024-09-30 14:08:19.364444] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:15.003 [2024-09-30 14:08:19.405749] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:15.003 [2024-09-30 14:08:19.405791] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:15.575 14:08:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 BaseBdev1_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 true 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 [2024-09-30 14:08:20.038585] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:15.575 [2024-09-30 14:08:20.038734] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:15.575 [2024-09-30 14:08:20.038767] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:15.575 [2024-09-30 14:08:20.038798] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:15.575 [2024-09-30 14:08:20.040832] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:15.575 [2024-09-30 14:08:20.040920] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:15.575 BaseBdev1 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 BaseBdev2_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 true 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 [2024-09-30 14:08:20.092283] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:15.575 [2024-09-30 14:08:20.092411] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:15.575 [2024-09-30 14:08:20.092460] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:15.575 [2024-09-30 14:08:20.092499] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:15.575 [2024-09-30 14:08:20.094485] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:15.575 [2024-09-30 14:08:20.094595] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:15.575 BaseBdev2 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 BaseBdev3_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 true 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 [2024-09-30 14:08:20.132452] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:15.575 [2024-09-30 14:08:20.132570] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:15.575 [2024-09-30 14:08:20.132602] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:15.575 [2024-09-30 14:08:20.132631] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:15.575 [2024-09-30 14:08:20.134542] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:15.575 [2024-09-30 14:08:20.134610] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:15.575 BaseBdev3 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 [2024-09-30 14:08:20.144491] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:15.575 [2024-09-30 14:08:20.146244] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:15.575 [2024-09-30 14:08:20.146344] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:15.575 [2024-09-30 14:08:20.146554] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:15.575 [2024-09-30 14:08:20.146598] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:15.575 [2024-09-30 14:08:20.146868] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:15.575 [2024-09-30 14:08:20.147058] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:15.575 [2024-09-30 14:08:20.147080] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:15.575 [2024-09-30 14:08:20.147191] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.575 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.576 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.576 "name": "raid_bdev1", 00:09:15.576 "uuid": "94ddac1d-2fe6-4df3-a1b0-d6f2e341dadd", 00:09:15.576 "strip_size_kb": 0, 00:09:15.576 "state": "online", 00:09:15.576 "raid_level": "raid1", 00:09:15.576 "superblock": true, 00:09:15.576 "num_base_bdevs": 3, 00:09:15.576 "num_base_bdevs_discovered": 3, 00:09:15.576 "num_base_bdevs_operational": 3, 00:09:15.576 "base_bdevs_list": [ 00:09:15.576 { 00:09:15.576 "name": "BaseBdev1", 00:09:15.576 "uuid": "68e3e996-a6d7-5d7b-b018-970b34a33483", 00:09:15.576 "is_configured": true, 00:09:15.576 "data_offset": 2048, 00:09:15.576 "data_size": 63488 00:09:15.576 }, 00:09:15.576 { 00:09:15.576 "name": "BaseBdev2", 00:09:15.576 "uuid": "042831a1-3210-5b34-8a5d-34360a2553d1", 00:09:15.576 "is_configured": true, 00:09:15.576 "data_offset": 2048, 00:09:15.576 "data_size": 63488 00:09:15.576 }, 00:09:15.576 { 00:09:15.576 "name": "BaseBdev3", 00:09:15.576 "uuid": "b2da94e9-ffb2-5686-adfc-eba404b3a78f", 00:09:15.576 "is_configured": true, 00:09:15.576 "data_offset": 2048, 00:09:15.576 "data_size": 63488 00:09:15.576 } 00:09:15.576 ] 00:09:15.576 }' 00:09:15.576 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.576 14:08:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.166 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:16.166 14:08:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:16.166 [2024-09-30 14:08:20.632067] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.105 "name": "raid_bdev1", 00:09:17.105 "uuid": "94ddac1d-2fe6-4df3-a1b0-d6f2e341dadd", 00:09:17.105 "strip_size_kb": 0, 00:09:17.105 "state": "online", 00:09:17.105 "raid_level": "raid1", 00:09:17.105 "superblock": true, 00:09:17.105 "num_base_bdevs": 3, 00:09:17.105 "num_base_bdevs_discovered": 3, 00:09:17.105 "num_base_bdevs_operational": 3, 00:09:17.105 "base_bdevs_list": [ 00:09:17.105 { 00:09:17.105 "name": "BaseBdev1", 00:09:17.105 "uuid": "68e3e996-a6d7-5d7b-b018-970b34a33483", 00:09:17.105 "is_configured": true, 00:09:17.105 "data_offset": 2048, 00:09:17.105 "data_size": 63488 00:09:17.105 }, 00:09:17.105 { 00:09:17.105 "name": "BaseBdev2", 00:09:17.105 "uuid": "042831a1-3210-5b34-8a5d-34360a2553d1", 00:09:17.105 "is_configured": true, 00:09:17.105 "data_offset": 2048, 00:09:17.105 "data_size": 63488 00:09:17.105 }, 00:09:17.105 { 00:09:17.105 "name": "BaseBdev3", 00:09:17.105 "uuid": "b2da94e9-ffb2-5686-adfc-eba404b3a78f", 00:09:17.105 "is_configured": true, 00:09:17.105 "data_offset": 2048, 00:09:17.105 "data_size": 63488 00:09:17.105 } 00:09:17.105 ] 00:09:17.105 }' 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.105 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.365 [2024-09-30 14:08:21.968632] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:17.365 [2024-09-30 14:08:21.968755] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:17.365 [2024-09-30 14:08:21.971134] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:17.365 [2024-09-30 14:08:21.971217] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:17.365 [2024-09-30 14:08:21.971337] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:17.365 [2024-09-30 14:08:21.971349] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.365 { 00:09:17.365 "results": [ 00:09:17.365 { 00:09:17.365 "job": "raid_bdev1", 00:09:17.365 "core_mask": "0x1", 00:09:17.365 "workload": "randrw", 00:09:17.365 "percentage": 50, 00:09:17.365 "status": "finished", 00:09:17.365 "queue_depth": 1, 00:09:17.365 "io_size": 131072, 00:09:17.365 "runtime": 1.337408, 00:09:17.365 "iops": 15310.211992151984, 00:09:17.365 "mibps": 1913.776499018998, 00:09:17.365 "io_failed": 0, 00:09:17.365 "io_timeout": 0, 00:09:17.365 "avg_latency_us": 62.901859755291305, 00:09:17.365 "min_latency_us": 21.687336244541484, 00:09:17.365 "max_latency_us": 1416.6078602620087 00:09:17.365 } 00:09:17.365 ], 00:09:17.365 "core_count": 1 00:09:17.365 } 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81127 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 81127 ']' 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 81127 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:17.365 14:08:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81127 00:09:17.365 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:17.365 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81127' 00:09:17.624 killing process with pid 81127 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 81127 00:09:17.624 [2024-09-30 14:08:22.021324] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 81127 00:09:17.624 [2024-09-30 14:08:22.046360] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4NgMFUbOgf 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:17.624 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:17.884 ************************************ 00:09:17.884 END TEST raid_read_error_test 00:09:17.884 ************************************ 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:17.884 00:09:17.884 real 0m3.216s 00:09:17.884 user 0m3.983s 00:09:17.884 sys 0m0.560s 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:17.884 14:08:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.884 14:08:22 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:17.884 14:08:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:17.884 14:08:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:17.884 14:08:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:17.884 ************************************ 00:09:17.884 START TEST raid_write_error_test 00:09:17.884 ************************************ 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 write 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:17.884 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.DjKjis3AZQ 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81260 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81260 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 81260 ']' 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:17.885 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:17.885 14:08:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.885 [2024-09-30 14:08:22.474587] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:17.885 [2024-09-30 14:08:22.474862] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81260 ] 00:09:18.145 [2024-09-30 14:08:22.612407] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:18.145 [2024-09-30 14:08:22.641330] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:18.145 [2024-09-30 14:08:22.687217] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:18.145 [2024-09-30 14:08:22.728105] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:18.145 [2024-09-30 14:08:22.728142] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.713 BaseBdev1_malloc 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.713 true 00:09:18.713 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.714 [2024-09-30 14:08:23.317827] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:18.714 [2024-09-30 14:08:23.317982] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.714 [2024-09-30 14:08:23.318020] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:18.714 [2024-09-30 14:08:23.318052] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.714 [2024-09-30 14:08:23.320155] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.714 [2024-09-30 14:08:23.320227] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:18.714 BaseBdev1 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.714 BaseBdev2_malloc 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.714 true 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.714 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.714 [2024-09-30 14:08:23.366745] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:18.714 [2024-09-30 14:08:23.366854] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.714 [2024-09-30 14:08:23.366888] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:18.714 [2024-09-30 14:08:23.366918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.714 [2024-09-30 14:08:23.368942] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.714 [2024-09-30 14:08:23.369021] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:18.974 BaseBdev2 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.974 BaseBdev3_malloc 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.974 true 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.974 [2024-09-30 14:08:23.407094] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:18.974 [2024-09-30 14:08:23.407200] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.974 [2024-09-30 14:08:23.407220] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:18.974 [2024-09-30 14:08:23.407232] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.974 [2024-09-30 14:08:23.409240] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.974 [2024-09-30 14:08:23.409312] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:18.974 BaseBdev3 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.974 [2024-09-30 14:08:23.419142] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.974 [2024-09-30 14:08:23.420963] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:18.974 [2024-09-30 14:08:23.421066] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:18.974 [2024-09-30 14:08:23.421259] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:18.974 [2024-09-30 14:08:23.421304] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:18.974 [2024-09-30 14:08:23.421552] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:18.974 [2024-09-30 14:08:23.421726] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:18.974 [2024-09-30 14:08:23.421771] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:18.974 [2024-09-30 14:08:23.421922] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.974 "name": "raid_bdev1", 00:09:18.974 "uuid": "ad62251b-4ac4-4a00-b637-1724840ee7e9", 00:09:18.974 "strip_size_kb": 0, 00:09:18.974 "state": "online", 00:09:18.974 "raid_level": "raid1", 00:09:18.974 "superblock": true, 00:09:18.974 "num_base_bdevs": 3, 00:09:18.974 "num_base_bdevs_discovered": 3, 00:09:18.974 "num_base_bdevs_operational": 3, 00:09:18.974 "base_bdevs_list": [ 00:09:18.974 { 00:09:18.974 "name": "BaseBdev1", 00:09:18.974 "uuid": "0e414012-333d-516b-9e05-40d9fad05694", 00:09:18.974 "is_configured": true, 00:09:18.974 "data_offset": 2048, 00:09:18.974 "data_size": 63488 00:09:18.974 }, 00:09:18.974 { 00:09:18.974 "name": "BaseBdev2", 00:09:18.974 "uuid": "8e4de70c-f8ac-587b-bd38-3fd0001f295f", 00:09:18.974 "is_configured": true, 00:09:18.974 "data_offset": 2048, 00:09:18.974 "data_size": 63488 00:09:18.974 }, 00:09:18.974 { 00:09:18.974 "name": "BaseBdev3", 00:09:18.974 "uuid": "b19b1e2a-29b6-573a-96b7-142b8f387bc8", 00:09:18.974 "is_configured": true, 00:09:18.974 "data_offset": 2048, 00:09:18.974 "data_size": 63488 00:09:18.974 } 00:09:18.974 ] 00:09:18.974 }' 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.974 14:08:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.234 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:19.234 14:08:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:19.494 [2024-09-30 14:08:23.954724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.435 [2024-09-30 14:08:24.877598] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:20.435 [2024-09-30 14:08:24.877763] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:20.435 [2024-09-30 14:08:24.878041] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.435 "name": "raid_bdev1", 00:09:20.435 "uuid": "ad62251b-4ac4-4a00-b637-1724840ee7e9", 00:09:20.435 "strip_size_kb": 0, 00:09:20.435 "state": "online", 00:09:20.435 "raid_level": "raid1", 00:09:20.435 "superblock": true, 00:09:20.435 "num_base_bdevs": 3, 00:09:20.435 "num_base_bdevs_discovered": 2, 00:09:20.435 "num_base_bdevs_operational": 2, 00:09:20.435 "base_bdevs_list": [ 00:09:20.435 { 00:09:20.435 "name": null, 00:09:20.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.435 "is_configured": false, 00:09:20.435 "data_offset": 0, 00:09:20.435 "data_size": 63488 00:09:20.435 }, 00:09:20.435 { 00:09:20.435 "name": "BaseBdev2", 00:09:20.435 "uuid": "8e4de70c-f8ac-587b-bd38-3fd0001f295f", 00:09:20.435 "is_configured": true, 00:09:20.435 "data_offset": 2048, 00:09:20.435 "data_size": 63488 00:09:20.435 }, 00:09:20.435 { 00:09:20.435 "name": "BaseBdev3", 00:09:20.435 "uuid": "b19b1e2a-29b6-573a-96b7-142b8f387bc8", 00:09:20.435 "is_configured": true, 00:09:20.435 "data_offset": 2048, 00:09:20.435 "data_size": 63488 00:09:20.435 } 00:09:20.435 ] 00:09:20.435 }' 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.435 14:08:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.696 [2024-09-30 14:08:25.283380] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:20.696 [2024-09-30 14:08:25.283537] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:20.696 [2024-09-30 14:08:25.285967] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:20.696 [2024-09-30 14:08:25.286048] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:20.696 [2024-09-30 14:08:25.286139] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:20.696 [2024-09-30 14:08:25.286188] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:20.696 { 00:09:20.696 "results": [ 00:09:20.696 { 00:09:20.696 "job": "raid_bdev1", 00:09:20.696 "core_mask": "0x1", 00:09:20.696 "workload": "randrw", 00:09:20.696 "percentage": 50, 00:09:20.696 "status": "finished", 00:09:20.696 "queue_depth": 1, 00:09:20.696 "io_size": 131072, 00:09:20.696 "runtime": 1.329558, 00:09:20.696 "iops": 16551.365190537006, 00:09:20.696 "mibps": 2068.9206488171258, 00:09:20.696 "io_failed": 0, 00:09:20.696 "io_timeout": 0, 00:09:20.696 "avg_latency_us": 57.92820584461483, 00:09:20.696 "min_latency_us": 21.575545851528386, 00:09:20.696 "max_latency_us": 1366.5257641921398 00:09:20.696 } 00:09:20.696 ], 00:09:20.696 "core_count": 1 00:09:20.696 } 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81260 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 81260 ']' 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 81260 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81260 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:20.696 killing process with pid 81260 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81260' 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 81260 00:09:20.696 [2024-09-30 14:08:25.323068] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:20.696 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 81260 00:09:20.696 [2024-09-30 14:08:25.348376] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.DjKjis3AZQ 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:20.957 00:09:20.957 real 0m3.229s 00:09:20.957 user 0m4.033s 00:09:20.957 sys 0m0.542s 00:09:20.957 ************************************ 00:09:20.957 END TEST raid_write_error_test 00:09:20.957 ************************************ 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:20.957 14:08:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.217 14:08:25 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:21.217 14:08:25 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:21.217 14:08:25 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:21.217 14:08:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:21.217 14:08:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:21.217 14:08:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:21.217 ************************************ 00:09:21.217 START TEST raid_state_function_test 00:09:21.217 ************************************ 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 false 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:21.217 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:21.218 Process raid pid: 81394 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81394 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81394' 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81394 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 81394 ']' 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:21.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:21.218 14:08:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.218 [2024-09-30 14:08:25.755494] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:21.218 [2024-09-30 14:08:25.755697] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:21.478 [2024-09-30 14:08:25.887300] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:21.478 [2024-09-30 14:08:25.914813] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:21.478 [2024-09-30 14:08:25.960228] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:21.478 [2024-09-30 14:08:26.001169] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:21.478 [2024-09-30 14:08:26.001271] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.046 [2024-09-30 14:08:26.590201] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:22.046 [2024-09-30 14:08:26.590301] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:22.046 [2024-09-30 14:08:26.590342] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:22.046 [2024-09-30 14:08:26.590364] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:22.046 [2024-09-30 14:08:26.590386] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:22.046 [2024-09-30 14:08:26.590404] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:22.046 [2024-09-30 14:08:26.590422] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:22.046 [2024-09-30 14:08:26.590440] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.046 "name": "Existed_Raid", 00:09:22.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.046 "strip_size_kb": 64, 00:09:22.046 "state": "configuring", 00:09:22.046 "raid_level": "raid0", 00:09:22.046 "superblock": false, 00:09:22.046 "num_base_bdevs": 4, 00:09:22.046 "num_base_bdevs_discovered": 0, 00:09:22.046 "num_base_bdevs_operational": 4, 00:09:22.046 "base_bdevs_list": [ 00:09:22.046 { 00:09:22.046 "name": "BaseBdev1", 00:09:22.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.046 "is_configured": false, 00:09:22.046 "data_offset": 0, 00:09:22.046 "data_size": 0 00:09:22.046 }, 00:09:22.046 { 00:09:22.046 "name": "BaseBdev2", 00:09:22.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.046 "is_configured": false, 00:09:22.046 "data_offset": 0, 00:09:22.046 "data_size": 0 00:09:22.046 }, 00:09:22.046 { 00:09:22.046 "name": "BaseBdev3", 00:09:22.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.046 "is_configured": false, 00:09:22.046 "data_offset": 0, 00:09:22.046 "data_size": 0 00:09:22.046 }, 00:09:22.046 { 00:09:22.046 "name": "BaseBdev4", 00:09:22.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.046 "is_configured": false, 00:09:22.046 "data_offset": 0, 00:09:22.046 "data_size": 0 00:09:22.046 } 00:09:22.046 ] 00:09:22.046 }' 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.046 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 14:08:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:22.615 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.615 14:08:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 [2024-09-30 14:08:27.005466] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:22.615 [2024-09-30 14:08:27.005584] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 [2024-09-30 14:08:27.017432] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:22.615 [2024-09-30 14:08:27.017527] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:22.615 [2024-09-30 14:08:27.017578] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:22.615 [2024-09-30 14:08:27.017598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:22.615 [2024-09-30 14:08:27.017617] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:22.615 [2024-09-30 14:08:27.017635] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:22.615 [2024-09-30 14:08:27.017654] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:22.615 [2024-09-30 14:08:27.017672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 [2024-09-30 14:08:27.037996] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:22.615 BaseBdev1 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.615 [ 00:09:22.615 { 00:09:22.615 "name": "BaseBdev1", 00:09:22.615 "aliases": [ 00:09:22.615 "3e8f459e-f64e-4c10-a781-525ccbb3743f" 00:09:22.615 ], 00:09:22.615 "product_name": "Malloc disk", 00:09:22.615 "block_size": 512, 00:09:22.615 "num_blocks": 65536, 00:09:22.615 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:22.615 "assigned_rate_limits": { 00:09:22.615 "rw_ios_per_sec": 0, 00:09:22.615 "rw_mbytes_per_sec": 0, 00:09:22.615 "r_mbytes_per_sec": 0, 00:09:22.615 "w_mbytes_per_sec": 0 00:09:22.615 }, 00:09:22.615 "claimed": true, 00:09:22.615 "claim_type": "exclusive_write", 00:09:22.615 "zoned": false, 00:09:22.615 "supported_io_types": { 00:09:22.615 "read": true, 00:09:22.615 "write": true, 00:09:22.615 "unmap": true, 00:09:22.615 "flush": true, 00:09:22.615 "reset": true, 00:09:22.615 "nvme_admin": false, 00:09:22.615 "nvme_io": false, 00:09:22.615 "nvme_io_md": false, 00:09:22.615 "write_zeroes": true, 00:09:22.615 "zcopy": true, 00:09:22.615 "get_zone_info": false, 00:09:22.615 "zone_management": false, 00:09:22.615 "zone_append": false, 00:09:22.615 "compare": false, 00:09:22.615 "compare_and_write": false, 00:09:22.615 "abort": true, 00:09:22.615 "seek_hole": false, 00:09:22.615 "seek_data": false, 00:09:22.615 "copy": true, 00:09:22.615 "nvme_iov_md": false 00:09:22.615 }, 00:09:22.615 "memory_domains": [ 00:09:22.615 { 00:09:22.615 "dma_device_id": "system", 00:09:22.615 "dma_device_type": 1 00:09:22.615 }, 00:09:22.615 { 00:09:22.615 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.615 "dma_device_type": 2 00:09:22.615 } 00:09:22.615 ], 00:09:22.615 "driver_specific": {} 00:09:22.615 } 00:09:22.615 ] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.615 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.616 "name": "Existed_Raid", 00:09:22.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.616 "strip_size_kb": 64, 00:09:22.616 "state": "configuring", 00:09:22.616 "raid_level": "raid0", 00:09:22.616 "superblock": false, 00:09:22.616 "num_base_bdevs": 4, 00:09:22.616 "num_base_bdevs_discovered": 1, 00:09:22.616 "num_base_bdevs_operational": 4, 00:09:22.616 "base_bdevs_list": [ 00:09:22.616 { 00:09:22.616 "name": "BaseBdev1", 00:09:22.616 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:22.616 "is_configured": true, 00:09:22.616 "data_offset": 0, 00:09:22.616 "data_size": 65536 00:09:22.616 }, 00:09:22.616 { 00:09:22.616 "name": "BaseBdev2", 00:09:22.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.616 "is_configured": false, 00:09:22.616 "data_offset": 0, 00:09:22.616 "data_size": 0 00:09:22.616 }, 00:09:22.616 { 00:09:22.616 "name": "BaseBdev3", 00:09:22.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.616 "is_configured": false, 00:09:22.616 "data_offset": 0, 00:09:22.616 "data_size": 0 00:09:22.616 }, 00:09:22.616 { 00:09:22.616 "name": "BaseBdev4", 00:09:22.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.616 "is_configured": false, 00:09:22.616 "data_offset": 0, 00:09:22.616 "data_size": 0 00:09:22.616 } 00:09:22.616 ] 00:09:22.616 }' 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.616 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.876 [2024-09-30 14:08:27.489255] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:22.876 [2024-09-30 14:08:27.489349] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.876 [2024-09-30 14:08:27.501263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:22.876 [2024-09-30 14:08:27.503103] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:22.876 [2024-09-30 14:08:27.503173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:22.876 [2024-09-30 14:08:27.503202] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:22.876 [2024-09-30 14:08:27.503222] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:22.876 [2024-09-30 14:08:27.503240] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:22.876 [2024-09-30 14:08:27.503258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.876 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.136 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.136 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.136 "name": "Existed_Raid", 00:09:23.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.136 "strip_size_kb": 64, 00:09:23.136 "state": "configuring", 00:09:23.136 "raid_level": "raid0", 00:09:23.136 "superblock": false, 00:09:23.136 "num_base_bdevs": 4, 00:09:23.136 "num_base_bdevs_discovered": 1, 00:09:23.136 "num_base_bdevs_operational": 4, 00:09:23.136 "base_bdevs_list": [ 00:09:23.136 { 00:09:23.136 "name": "BaseBdev1", 00:09:23.136 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:23.136 "is_configured": true, 00:09:23.136 "data_offset": 0, 00:09:23.137 "data_size": 65536 00:09:23.137 }, 00:09:23.137 { 00:09:23.137 "name": "BaseBdev2", 00:09:23.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.137 "is_configured": false, 00:09:23.137 "data_offset": 0, 00:09:23.137 "data_size": 0 00:09:23.137 }, 00:09:23.137 { 00:09:23.137 "name": "BaseBdev3", 00:09:23.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.137 "is_configured": false, 00:09:23.137 "data_offset": 0, 00:09:23.137 "data_size": 0 00:09:23.137 }, 00:09:23.137 { 00:09:23.137 "name": "BaseBdev4", 00:09:23.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.137 "is_configured": false, 00:09:23.137 "data_offset": 0, 00:09:23.137 "data_size": 0 00:09:23.137 } 00:09:23.137 ] 00:09:23.137 }' 00:09:23.137 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.137 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.397 [2024-09-30 14:08:27.982530] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:23.397 BaseBdev2 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.397 14:08:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.397 [ 00:09:23.397 { 00:09:23.397 "name": "BaseBdev2", 00:09:23.397 "aliases": [ 00:09:23.397 "4e42aeca-af7e-4bfe-ad4c-755440986f33" 00:09:23.397 ], 00:09:23.397 "product_name": "Malloc disk", 00:09:23.397 "block_size": 512, 00:09:23.397 "num_blocks": 65536, 00:09:23.397 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:23.397 "assigned_rate_limits": { 00:09:23.397 "rw_ios_per_sec": 0, 00:09:23.397 "rw_mbytes_per_sec": 0, 00:09:23.397 "r_mbytes_per_sec": 0, 00:09:23.397 "w_mbytes_per_sec": 0 00:09:23.397 }, 00:09:23.397 "claimed": true, 00:09:23.397 "claim_type": "exclusive_write", 00:09:23.397 "zoned": false, 00:09:23.397 "supported_io_types": { 00:09:23.397 "read": true, 00:09:23.397 "write": true, 00:09:23.397 "unmap": true, 00:09:23.397 "flush": true, 00:09:23.397 "reset": true, 00:09:23.397 "nvme_admin": false, 00:09:23.397 "nvme_io": false, 00:09:23.397 "nvme_io_md": false, 00:09:23.397 "write_zeroes": true, 00:09:23.397 "zcopy": true, 00:09:23.397 "get_zone_info": false, 00:09:23.397 "zone_management": false, 00:09:23.397 "zone_append": false, 00:09:23.397 "compare": false, 00:09:23.397 "compare_and_write": false, 00:09:23.397 "abort": true, 00:09:23.397 "seek_hole": false, 00:09:23.397 "seek_data": false, 00:09:23.397 "copy": true, 00:09:23.397 "nvme_iov_md": false 00:09:23.397 }, 00:09:23.397 "memory_domains": [ 00:09:23.397 { 00:09:23.397 "dma_device_id": "system", 00:09:23.397 "dma_device_type": 1 00:09:23.397 }, 00:09:23.397 { 00:09:23.397 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.397 "dma_device_type": 2 00:09:23.397 } 00:09:23.397 ], 00:09:23.397 "driver_specific": {} 00:09:23.397 } 00:09:23.397 ] 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.397 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.657 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.657 "name": "Existed_Raid", 00:09:23.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.657 "strip_size_kb": 64, 00:09:23.657 "state": "configuring", 00:09:23.657 "raid_level": "raid0", 00:09:23.657 "superblock": false, 00:09:23.657 "num_base_bdevs": 4, 00:09:23.657 "num_base_bdevs_discovered": 2, 00:09:23.657 "num_base_bdevs_operational": 4, 00:09:23.657 "base_bdevs_list": [ 00:09:23.657 { 00:09:23.657 "name": "BaseBdev1", 00:09:23.657 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:23.657 "is_configured": true, 00:09:23.657 "data_offset": 0, 00:09:23.657 "data_size": 65536 00:09:23.657 }, 00:09:23.657 { 00:09:23.657 "name": "BaseBdev2", 00:09:23.657 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:23.657 "is_configured": true, 00:09:23.657 "data_offset": 0, 00:09:23.657 "data_size": 65536 00:09:23.657 }, 00:09:23.657 { 00:09:23.657 "name": "BaseBdev3", 00:09:23.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.657 "is_configured": false, 00:09:23.657 "data_offset": 0, 00:09:23.657 "data_size": 0 00:09:23.657 }, 00:09:23.657 { 00:09:23.657 "name": "BaseBdev4", 00:09:23.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.657 "is_configured": false, 00:09:23.657 "data_offset": 0, 00:09:23.657 "data_size": 0 00:09:23.657 } 00:09:23.657 ] 00:09:23.657 }' 00:09:23.657 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.657 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.917 [2024-09-30 14:08:28.440480] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:23.917 BaseBdev3 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.917 [ 00:09:23.917 { 00:09:23.917 "name": "BaseBdev3", 00:09:23.917 "aliases": [ 00:09:23.917 "8f5e5008-b029-47e8-902c-ef587da447fe" 00:09:23.917 ], 00:09:23.917 "product_name": "Malloc disk", 00:09:23.917 "block_size": 512, 00:09:23.917 "num_blocks": 65536, 00:09:23.917 "uuid": "8f5e5008-b029-47e8-902c-ef587da447fe", 00:09:23.917 "assigned_rate_limits": { 00:09:23.917 "rw_ios_per_sec": 0, 00:09:23.917 "rw_mbytes_per_sec": 0, 00:09:23.917 "r_mbytes_per_sec": 0, 00:09:23.917 "w_mbytes_per_sec": 0 00:09:23.917 }, 00:09:23.917 "claimed": true, 00:09:23.917 "claim_type": "exclusive_write", 00:09:23.917 "zoned": false, 00:09:23.917 "supported_io_types": { 00:09:23.917 "read": true, 00:09:23.917 "write": true, 00:09:23.917 "unmap": true, 00:09:23.917 "flush": true, 00:09:23.917 "reset": true, 00:09:23.917 "nvme_admin": false, 00:09:23.917 "nvme_io": false, 00:09:23.917 "nvme_io_md": false, 00:09:23.917 "write_zeroes": true, 00:09:23.917 "zcopy": true, 00:09:23.917 "get_zone_info": false, 00:09:23.917 "zone_management": false, 00:09:23.917 "zone_append": false, 00:09:23.917 "compare": false, 00:09:23.917 "compare_and_write": false, 00:09:23.917 "abort": true, 00:09:23.917 "seek_hole": false, 00:09:23.917 "seek_data": false, 00:09:23.917 "copy": true, 00:09:23.917 "nvme_iov_md": false 00:09:23.917 }, 00:09:23.917 "memory_domains": [ 00:09:23.917 { 00:09:23.917 "dma_device_id": "system", 00:09:23.917 "dma_device_type": 1 00:09:23.917 }, 00:09:23.917 { 00:09:23.917 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.917 "dma_device_type": 2 00:09:23.917 } 00:09:23.917 ], 00:09:23.917 "driver_specific": {} 00:09:23.917 } 00:09:23.917 ] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.917 "name": "Existed_Raid", 00:09:23.917 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.917 "strip_size_kb": 64, 00:09:23.917 "state": "configuring", 00:09:23.917 "raid_level": "raid0", 00:09:23.917 "superblock": false, 00:09:23.917 "num_base_bdevs": 4, 00:09:23.917 "num_base_bdevs_discovered": 3, 00:09:23.917 "num_base_bdevs_operational": 4, 00:09:23.917 "base_bdevs_list": [ 00:09:23.917 { 00:09:23.917 "name": "BaseBdev1", 00:09:23.917 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:23.917 "is_configured": true, 00:09:23.917 "data_offset": 0, 00:09:23.917 "data_size": 65536 00:09:23.917 }, 00:09:23.917 { 00:09:23.917 "name": "BaseBdev2", 00:09:23.917 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:23.917 "is_configured": true, 00:09:23.917 "data_offset": 0, 00:09:23.917 "data_size": 65536 00:09:23.917 }, 00:09:23.917 { 00:09:23.917 "name": "BaseBdev3", 00:09:23.917 "uuid": "8f5e5008-b029-47e8-902c-ef587da447fe", 00:09:23.917 "is_configured": true, 00:09:23.917 "data_offset": 0, 00:09:23.917 "data_size": 65536 00:09:23.917 }, 00:09:23.917 { 00:09:23.917 "name": "BaseBdev4", 00:09:23.917 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.917 "is_configured": false, 00:09:23.917 "data_offset": 0, 00:09:23.917 "data_size": 0 00:09:23.917 } 00:09:23.917 ] 00:09:23.917 }' 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.917 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.488 [2024-09-30 14:08:28.934375] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:24.488 [2024-09-30 14:08:28.934489] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:24.488 [2024-09-30 14:08:28.934519] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:24.488 [2024-09-30 14:08:28.934844] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:24.488 BaseBdev4 00:09:24.488 [2024-09-30 14:08:28.935011] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:24.488 [2024-09-30 14:08:28.935035] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:24.488 [2024-09-30 14:08:28.935246] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.488 [ 00:09:24.488 { 00:09:24.488 "name": "BaseBdev4", 00:09:24.488 "aliases": [ 00:09:24.488 "416ff570-8179-41f6-8094-74ace077c7d7" 00:09:24.488 ], 00:09:24.488 "product_name": "Malloc disk", 00:09:24.488 "block_size": 512, 00:09:24.488 "num_blocks": 65536, 00:09:24.488 "uuid": "416ff570-8179-41f6-8094-74ace077c7d7", 00:09:24.488 "assigned_rate_limits": { 00:09:24.488 "rw_ios_per_sec": 0, 00:09:24.488 "rw_mbytes_per_sec": 0, 00:09:24.488 "r_mbytes_per_sec": 0, 00:09:24.488 "w_mbytes_per_sec": 0 00:09:24.488 }, 00:09:24.488 "claimed": true, 00:09:24.488 "claim_type": "exclusive_write", 00:09:24.488 "zoned": false, 00:09:24.488 "supported_io_types": { 00:09:24.488 "read": true, 00:09:24.488 "write": true, 00:09:24.488 "unmap": true, 00:09:24.488 "flush": true, 00:09:24.488 "reset": true, 00:09:24.488 "nvme_admin": false, 00:09:24.488 "nvme_io": false, 00:09:24.488 "nvme_io_md": false, 00:09:24.488 "write_zeroes": true, 00:09:24.488 "zcopy": true, 00:09:24.488 "get_zone_info": false, 00:09:24.488 "zone_management": false, 00:09:24.488 "zone_append": false, 00:09:24.488 "compare": false, 00:09:24.488 "compare_and_write": false, 00:09:24.488 "abort": true, 00:09:24.488 "seek_hole": false, 00:09:24.488 "seek_data": false, 00:09:24.488 "copy": true, 00:09:24.488 "nvme_iov_md": false 00:09:24.488 }, 00:09:24.488 "memory_domains": [ 00:09:24.488 { 00:09:24.488 "dma_device_id": "system", 00:09:24.488 "dma_device_type": 1 00:09:24.488 }, 00:09:24.488 { 00:09:24.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.488 "dma_device_type": 2 00:09:24.488 } 00:09:24.488 ], 00:09:24.488 "driver_specific": {} 00:09:24.488 } 00:09:24.488 ] 00:09:24.488 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.489 14:08:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.489 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.489 "name": "Existed_Raid", 00:09:24.489 "uuid": "022d2af0-f777-4cd6-a98d-faeaec746e7b", 00:09:24.489 "strip_size_kb": 64, 00:09:24.489 "state": "online", 00:09:24.489 "raid_level": "raid0", 00:09:24.489 "superblock": false, 00:09:24.489 "num_base_bdevs": 4, 00:09:24.489 "num_base_bdevs_discovered": 4, 00:09:24.489 "num_base_bdevs_operational": 4, 00:09:24.489 "base_bdevs_list": [ 00:09:24.489 { 00:09:24.489 "name": "BaseBdev1", 00:09:24.489 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:24.489 "is_configured": true, 00:09:24.489 "data_offset": 0, 00:09:24.489 "data_size": 65536 00:09:24.489 }, 00:09:24.489 { 00:09:24.489 "name": "BaseBdev2", 00:09:24.489 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:24.489 "is_configured": true, 00:09:24.489 "data_offset": 0, 00:09:24.489 "data_size": 65536 00:09:24.489 }, 00:09:24.489 { 00:09:24.489 "name": "BaseBdev3", 00:09:24.489 "uuid": "8f5e5008-b029-47e8-902c-ef587da447fe", 00:09:24.489 "is_configured": true, 00:09:24.489 "data_offset": 0, 00:09:24.489 "data_size": 65536 00:09:24.489 }, 00:09:24.489 { 00:09:24.489 "name": "BaseBdev4", 00:09:24.489 "uuid": "416ff570-8179-41f6-8094-74ace077c7d7", 00:09:24.489 "is_configured": true, 00:09:24.489 "data_offset": 0, 00:09:24.489 "data_size": 65536 00:09:24.489 } 00:09:24.489 ] 00:09:24.489 }' 00:09:24.489 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.489 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:24.750 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:25.010 [2024-09-30 14:08:29.409919] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:25.010 "name": "Existed_Raid", 00:09:25.010 "aliases": [ 00:09:25.010 "022d2af0-f777-4cd6-a98d-faeaec746e7b" 00:09:25.010 ], 00:09:25.010 "product_name": "Raid Volume", 00:09:25.010 "block_size": 512, 00:09:25.010 "num_blocks": 262144, 00:09:25.010 "uuid": "022d2af0-f777-4cd6-a98d-faeaec746e7b", 00:09:25.010 "assigned_rate_limits": { 00:09:25.010 "rw_ios_per_sec": 0, 00:09:25.010 "rw_mbytes_per_sec": 0, 00:09:25.010 "r_mbytes_per_sec": 0, 00:09:25.010 "w_mbytes_per_sec": 0 00:09:25.010 }, 00:09:25.010 "claimed": false, 00:09:25.010 "zoned": false, 00:09:25.010 "supported_io_types": { 00:09:25.010 "read": true, 00:09:25.010 "write": true, 00:09:25.010 "unmap": true, 00:09:25.010 "flush": true, 00:09:25.010 "reset": true, 00:09:25.010 "nvme_admin": false, 00:09:25.010 "nvme_io": false, 00:09:25.010 "nvme_io_md": false, 00:09:25.010 "write_zeroes": true, 00:09:25.010 "zcopy": false, 00:09:25.010 "get_zone_info": false, 00:09:25.010 "zone_management": false, 00:09:25.010 "zone_append": false, 00:09:25.010 "compare": false, 00:09:25.010 "compare_and_write": false, 00:09:25.010 "abort": false, 00:09:25.010 "seek_hole": false, 00:09:25.010 "seek_data": false, 00:09:25.010 "copy": false, 00:09:25.010 "nvme_iov_md": false 00:09:25.010 }, 00:09:25.010 "memory_domains": [ 00:09:25.010 { 00:09:25.010 "dma_device_id": "system", 00:09:25.010 "dma_device_type": 1 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.010 "dma_device_type": 2 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "system", 00:09:25.010 "dma_device_type": 1 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.010 "dma_device_type": 2 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "system", 00:09:25.010 "dma_device_type": 1 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.010 "dma_device_type": 2 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "system", 00:09:25.010 "dma_device_type": 1 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.010 "dma_device_type": 2 00:09:25.010 } 00:09:25.010 ], 00:09:25.010 "driver_specific": { 00:09:25.010 "raid": { 00:09:25.010 "uuid": "022d2af0-f777-4cd6-a98d-faeaec746e7b", 00:09:25.010 "strip_size_kb": 64, 00:09:25.010 "state": "online", 00:09:25.010 "raid_level": "raid0", 00:09:25.010 "superblock": false, 00:09:25.010 "num_base_bdevs": 4, 00:09:25.010 "num_base_bdevs_discovered": 4, 00:09:25.010 "num_base_bdevs_operational": 4, 00:09:25.010 "base_bdevs_list": [ 00:09:25.010 { 00:09:25.010 "name": "BaseBdev1", 00:09:25.010 "uuid": "3e8f459e-f64e-4c10-a781-525ccbb3743f", 00:09:25.010 "is_configured": true, 00:09:25.010 "data_offset": 0, 00:09:25.010 "data_size": 65536 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "name": "BaseBdev2", 00:09:25.010 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:25.010 "is_configured": true, 00:09:25.010 "data_offset": 0, 00:09:25.010 "data_size": 65536 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "name": "BaseBdev3", 00:09:25.010 "uuid": "8f5e5008-b029-47e8-902c-ef587da447fe", 00:09:25.010 "is_configured": true, 00:09:25.010 "data_offset": 0, 00:09:25.010 "data_size": 65536 00:09:25.010 }, 00:09:25.010 { 00:09:25.010 "name": "BaseBdev4", 00:09:25.010 "uuid": "416ff570-8179-41f6-8094-74ace077c7d7", 00:09:25.010 "is_configured": true, 00:09:25.010 "data_offset": 0, 00:09:25.010 "data_size": 65536 00:09:25.010 } 00:09:25.010 ] 00:09:25.010 } 00:09:25.010 } 00:09:25.010 }' 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:25.010 BaseBdev2 00:09:25.010 BaseBdev3 00:09:25.010 BaseBdev4' 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:25.010 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.011 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.271 [2024-09-30 14:08:29.733073] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:25.271 [2024-09-30 14:08:29.733143] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:25.271 [2024-09-30 14:08:29.733223] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.271 "name": "Existed_Raid", 00:09:25.271 "uuid": "022d2af0-f777-4cd6-a98d-faeaec746e7b", 00:09:25.271 "strip_size_kb": 64, 00:09:25.271 "state": "offline", 00:09:25.271 "raid_level": "raid0", 00:09:25.271 "superblock": false, 00:09:25.271 "num_base_bdevs": 4, 00:09:25.271 "num_base_bdevs_discovered": 3, 00:09:25.271 "num_base_bdevs_operational": 3, 00:09:25.271 "base_bdevs_list": [ 00:09:25.271 { 00:09:25.271 "name": null, 00:09:25.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.271 "is_configured": false, 00:09:25.271 "data_offset": 0, 00:09:25.271 "data_size": 65536 00:09:25.271 }, 00:09:25.271 { 00:09:25.271 "name": "BaseBdev2", 00:09:25.271 "uuid": "4e42aeca-af7e-4bfe-ad4c-755440986f33", 00:09:25.271 "is_configured": true, 00:09:25.271 "data_offset": 0, 00:09:25.271 "data_size": 65536 00:09:25.271 }, 00:09:25.271 { 00:09:25.271 "name": "BaseBdev3", 00:09:25.271 "uuid": "8f5e5008-b029-47e8-902c-ef587da447fe", 00:09:25.271 "is_configured": true, 00:09:25.271 "data_offset": 0, 00:09:25.271 "data_size": 65536 00:09:25.271 }, 00:09:25.271 { 00:09:25.271 "name": "BaseBdev4", 00:09:25.271 "uuid": "416ff570-8179-41f6-8094-74ace077c7d7", 00:09:25.271 "is_configured": true, 00:09:25.271 "data_offset": 0, 00:09:25.271 "data_size": 65536 00:09:25.271 } 00:09:25.271 ] 00:09:25.271 }' 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.271 14:08:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:25.842 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 [2024-09-30 14:08:30.235159] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 [2024-09-30 14:08:30.305969] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 [2024-09-30 14:08:30.352878] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:25.843 [2024-09-30 14:08:30.352977] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 [ 00:09:25.843 { 00:09:25.843 "name": "BaseBdev2", 00:09:25.843 "aliases": [ 00:09:25.843 "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c" 00:09:25.843 ], 00:09:25.843 "product_name": "Malloc disk", 00:09:25.843 "block_size": 512, 00:09:25.843 "num_blocks": 65536, 00:09:25.843 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:25.843 "assigned_rate_limits": { 00:09:25.843 "rw_ios_per_sec": 0, 00:09:25.843 "rw_mbytes_per_sec": 0, 00:09:25.843 "r_mbytes_per_sec": 0, 00:09:25.843 "w_mbytes_per_sec": 0 00:09:25.843 }, 00:09:25.843 "claimed": false, 00:09:25.843 "zoned": false, 00:09:25.843 "supported_io_types": { 00:09:25.843 "read": true, 00:09:25.843 "write": true, 00:09:25.843 "unmap": true, 00:09:25.843 "flush": true, 00:09:25.843 "reset": true, 00:09:25.843 "nvme_admin": false, 00:09:25.843 "nvme_io": false, 00:09:25.843 "nvme_io_md": false, 00:09:25.843 "write_zeroes": true, 00:09:25.843 "zcopy": true, 00:09:25.843 "get_zone_info": false, 00:09:25.843 "zone_management": false, 00:09:25.843 "zone_append": false, 00:09:25.843 "compare": false, 00:09:25.843 "compare_and_write": false, 00:09:25.843 "abort": true, 00:09:25.843 "seek_hole": false, 00:09:25.843 "seek_data": false, 00:09:25.843 "copy": true, 00:09:25.843 "nvme_iov_md": false 00:09:25.843 }, 00:09:25.843 "memory_domains": [ 00:09:25.843 { 00:09:25.843 "dma_device_id": "system", 00:09:25.843 "dma_device_type": 1 00:09:25.843 }, 00:09:25.843 { 00:09:25.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.843 "dma_device_type": 2 00:09:25.843 } 00:09:25.843 ], 00:09:25.843 "driver_specific": {} 00:09:25.843 } 00:09:25.843 ] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.843 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.843 [ 00:09:25.843 { 00:09:25.843 "name": "BaseBdev3", 00:09:25.843 "aliases": [ 00:09:25.843 "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91" 00:09:25.843 ], 00:09:25.843 "product_name": "Malloc disk", 00:09:25.843 "block_size": 512, 00:09:25.843 "num_blocks": 65536, 00:09:25.843 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:25.843 "assigned_rate_limits": { 00:09:25.843 "rw_ios_per_sec": 0, 00:09:25.843 "rw_mbytes_per_sec": 0, 00:09:25.843 "r_mbytes_per_sec": 0, 00:09:25.843 "w_mbytes_per_sec": 0 00:09:25.843 }, 00:09:25.843 "claimed": false, 00:09:25.843 "zoned": false, 00:09:25.843 "supported_io_types": { 00:09:25.843 "read": true, 00:09:25.843 "write": true, 00:09:25.843 "unmap": true, 00:09:25.843 "flush": true, 00:09:25.843 "reset": true, 00:09:25.843 "nvme_admin": false, 00:09:25.844 "nvme_io": false, 00:09:26.104 "nvme_io_md": false, 00:09:26.104 "write_zeroes": true, 00:09:26.104 "zcopy": true, 00:09:26.104 "get_zone_info": false, 00:09:26.104 "zone_management": false, 00:09:26.104 "zone_append": false, 00:09:26.104 "compare": false, 00:09:26.104 "compare_and_write": false, 00:09:26.104 "abort": true, 00:09:26.104 "seek_hole": false, 00:09:26.104 "seek_data": false, 00:09:26.104 "copy": true, 00:09:26.104 "nvme_iov_md": false 00:09:26.104 }, 00:09:26.104 "memory_domains": [ 00:09:26.104 { 00:09:26.104 "dma_device_id": "system", 00:09:26.104 "dma_device_type": 1 00:09:26.104 }, 00:09:26.104 { 00:09:26.104 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.104 "dma_device_type": 2 00:09:26.104 } 00:09:26.104 ], 00:09:26.104 "driver_specific": {} 00:09:26.104 } 00:09:26.104 ] 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:26.104 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.105 BaseBdev4 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.105 [ 00:09:26.105 { 00:09:26.105 "name": "BaseBdev4", 00:09:26.105 "aliases": [ 00:09:26.105 "b6e0b777-9448-4934-8576-74e140ff430a" 00:09:26.105 ], 00:09:26.105 "product_name": "Malloc disk", 00:09:26.105 "block_size": 512, 00:09:26.105 "num_blocks": 65536, 00:09:26.105 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:26.105 "assigned_rate_limits": { 00:09:26.105 "rw_ios_per_sec": 0, 00:09:26.105 "rw_mbytes_per_sec": 0, 00:09:26.105 "r_mbytes_per_sec": 0, 00:09:26.105 "w_mbytes_per_sec": 0 00:09:26.105 }, 00:09:26.105 "claimed": false, 00:09:26.105 "zoned": false, 00:09:26.105 "supported_io_types": { 00:09:26.105 "read": true, 00:09:26.105 "write": true, 00:09:26.105 "unmap": true, 00:09:26.105 "flush": true, 00:09:26.105 "reset": true, 00:09:26.105 "nvme_admin": false, 00:09:26.105 "nvme_io": false, 00:09:26.105 "nvme_io_md": false, 00:09:26.105 "write_zeroes": true, 00:09:26.105 "zcopy": true, 00:09:26.105 "get_zone_info": false, 00:09:26.105 "zone_management": false, 00:09:26.105 "zone_append": false, 00:09:26.105 "compare": false, 00:09:26.105 "compare_and_write": false, 00:09:26.105 "abort": true, 00:09:26.105 "seek_hole": false, 00:09:26.105 "seek_data": false, 00:09:26.105 "copy": true, 00:09:26.105 "nvme_iov_md": false 00:09:26.105 }, 00:09:26.105 "memory_domains": [ 00:09:26.105 { 00:09:26.105 "dma_device_id": "system", 00:09:26.105 "dma_device_type": 1 00:09:26.105 }, 00:09:26.105 { 00:09:26.105 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.105 "dma_device_type": 2 00:09:26.105 } 00:09:26.105 ], 00:09:26.105 "driver_specific": {} 00:09:26.105 } 00:09:26.105 ] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.105 [2024-09-30 14:08:30.564227] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:26.105 [2024-09-30 14:08:30.564367] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:26.105 [2024-09-30 14:08:30.564404] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:26.105 [2024-09-30 14:08:30.566171] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:26.105 [2024-09-30 14:08:30.566257] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.105 "name": "Existed_Raid", 00:09:26.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.105 "strip_size_kb": 64, 00:09:26.105 "state": "configuring", 00:09:26.105 "raid_level": "raid0", 00:09:26.105 "superblock": false, 00:09:26.105 "num_base_bdevs": 4, 00:09:26.105 "num_base_bdevs_discovered": 3, 00:09:26.105 "num_base_bdevs_operational": 4, 00:09:26.105 "base_bdevs_list": [ 00:09:26.105 { 00:09:26.105 "name": "BaseBdev1", 00:09:26.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.105 "is_configured": false, 00:09:26.105 "data_offset": 0, 00:09:26.105 "data_size": 0 00:09:26.105 }, 00:09:26.105 { 00:09:26.105 "name": "BaseBdev2", 00:09:26.105 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:26.105 "is_configured": true, 00:09:26.105 "data_offset": 0, 00:09:26.105 "data_size": 65536 00:09:26.105 }, 00:09:26.105 { 00:09:26.105 "name": "BaseBdev3", 00:09:26.105 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:26.105 "is_configured": true, 00:09:26.105 "data_offset": 0, 00:09:26.105 "data_size": 65536 00:09:26.105 }, 00:09:26.105 { 00:09:26.105 "name": "BaseBdev4", 00:09:26.105 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:26.105 "is_configured": true, 00:09:26.105 "data_offset": 0, 00:09:26.105 "data_size": 65536 00:09:26.105 } 00:09:26.105 ] 00:09:26.105 }' 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.105 14:08:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.365 [2024-09-30 14:08:31.011474] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.365 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.625 "name": "Existed_Raid", 00:09:26.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.625 "strip_size_kb": 64, 00:09:26.625 "state": "configuring", 00:09:26.625 "raid_level": "raid0", 00:09:26.625 "superblock": false, 00:09:26.625 "num_base_bdevs": 4, 00:09:26.625 "num_base_bdevs_discovered": 2, 00:09:26.625 "num_base_bdevs_operational": 4, 00:09:26.625 "base_bdevs_list": [ 00:09:26.625 { 00:09:26.625 "name": "BaseBdev1", 00:09:26.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.625 "is_configured": false, 00:09:26.625 "data_offset": 0, 00:09:26.625 "data_size": 0 00:09:26.625 }, 00:09:26.625 { 00:09:26.625 "name": null, 00:09:26.625 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:26.625 "is_configured": false, 00:09:26.625 "data_offset": 0, 00:09:26.625 "data_size": 65536 00:09:26.625 }, 00:09:26.625 { 00:09:26.625 "name": "BaseBdev3", 00:09:26.625 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:26.625 "is_configured": true, 00:09:26.625 "data_offset": 0, 00:09:26.625 "data_size": 65536 00:09:26.625 }, 00:09:26.625 { 00:09:26.625 "name": "BaseBdev4", 00:09:26.625 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:26.625 "is_configured": true, 00:09:26.625 "data_offset": 0, 00:09:26.625 "data_size": 65536 00:09:26.625 } 00:09:26.625 ] 00:09:26.625 }' 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.625 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.885 [2024-09-30 14:08:31.481481] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:26.885 BaseBdev1 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.885 [ 00:09:26.885 { 00:09:26.885 "name": "BaseBdev1", 00:09:26.885 "aliases": [ 00:09:26.885 "46fab4c2-b38f-4890-9bec-55322e39e08e" 00:09:26.885 ], 00:09:26.885 "product_name": "Malloc disk", 00:09:26.885 "block_size": 512, 00:09:26.885 "num_blocks": 65536, 00:09:26.885 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:26.885 "assigned_rate_limits": { 00:09:26.885 "rw_ios_per_sec": 0, 00:09:26.885 "rw_mbytes_per_sec": 0, 00:09:26.885 "r_mbytes_per_sec": 0, 00:09:26.885 "w_mbytes_per_sec": 0 00:09:26.885 }, 00:09:26.885 "claimed": true, 00:09:26.885 "claim_type": "exclusive_write", 00:09:26.885 "zoned": false, 00:09:26.885 "supported_io_types": { 00:09:26.885 "read": true, 00:09:26.885 "write": true, 00:09:26.885 "unmap": true, 00:09:26.885 "flush": true, 00:09:26.885 "reset": true, 00:09:26.885 "nvme_admin": false, 00:09:26.885 "nvme_io": false, 00:09:26.885 "nvme_io_md": false, 00:09:26.885 "write_zeroes": true, 00:09:26.885 "zcopy": true, 00:09:26.885 "get_zone_info": false, 00:09:26.885 "zone_management": false, 00:09:26.885 "zone_append": false, 00:09:26.885 "compare": false, 00:09:26.885 "compare_and_write": false, 00:09:26.885 "abort": true, 00:09:26.885 "seek_hole": false, 00:09:26.885 "seek_data": false, 00:09:26.885 "copy": true, 00:09:26.885 "nvme_iov_md": false 00:09:26.885 }, 00:09:26.885 "memory_domains": [ 00:09:26.885 { 00:09:26.885 "dma_device_id": "system", 00:09:26.885 "dma_device_type": 1 00:09:26.885 }, 00:09:26.885 { 00:09:26.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.885 "dma_device_type": 2 00:09:26.885 } 00:09:26.885 ], 00:09:26.885 "driver_specific": {} 00:09:26.885 } 00:09:26.885 ] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.885 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.886 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.886 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.886 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.145 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.145 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.145 "name": "Existed_Raid", 00:09:27.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.145 "strip_size_kb": 64, 00:09:27.145 "state": "configuring", 00:09:27.145 "raid_level": "raid0", 00:09:27.145 "superblock": false, 00:09:27.145 "num_base_bdevs": 4, 00:09:27.145 "num_base_bdevs_discovered": 3, 00:09:27.145 "num_base_bdevs_operational": 4, 00:09:27.145 "base_bdevs_list": [ 00:09:27.145 { 00:09:27.145 "name": "BaseBdev1", 00:09:27.145 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:27.145 "is_configured": true, 00:09:27.145 "data_offset": 0, 00:09:27.145 "data_size": 65536 00:09:27.145 }, 00:09:27.145 { 00:09:27.145 "name": null, 00:09:27.145 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:27.145 "is_configured": false, 00:09:27.145 "data_offset": 0, 00:09:27.145 "data_size": 65536 00:09:27.145 }, 00:09:27.145 { 00:09:27.145 "name": "BaseBdev3", 00:09:27.145 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:27.145 "is_configured": true, 00:09:27.145 "data_offset": 0, 00:09:27.145 "data_size": 65536 00:09:27.145 }, 00:09:27.145 { 00:09:27.145 "name": "BaseBdev4", 00:09:27.145 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:27.145 "is_configured": true, 00:09:27.145 "data_offset": 0, 00:09:27.145 "data_size": 65536 00:09:27.145 } 00:09:27.145 ] 00:09:27.145 }' 00:09:27.145 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.145 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.405 [2024-09-30 14:08:31.984666] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.405 14:08:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.405 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.405 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.405 "name": "Existed_Raid", 00:09:27.405 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.405 "strip_size_kb": 64, 00:09:27.405 "state": "configuring", 00:09:27.405 "raid_level": "raid0", 00:09:27.405 "superblock": false, 00:09:27.405 "num_base_bdevs": 4, 00:09:27.405 "num_base_bdevs_discovered": 2, 00:09:27.405 "num_base_bdevs_operational": 4, 00:09:27.405 "base_bdevs_list": [ 00:09:27.405 { 00:09:27.405 "name": "BaseBdev1", 00:09:27.405 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:27.405 "is_configured": true, 00:09:27.405 "data_offset": 0, 00:09:27.405 "data_size": 65536 00:09:27.405 }, 00:09:27.405 { 00:09:27.405 "name": null, 00:09:27.405 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:27.405 "is_configured": false, 00:09:27.405 "data_offset": 0, 00:09:27.405 "data_size": 65536 00:09:27.405 }, 00:09:27.405 { 00:09:27.405 "name": null, 00:09:27.405 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:27.405 "is_configured": false, 00:09:27.405 "data_offset": 0, 00:09:27.405 "data_size": 65536 00:09:27.405 }, 00:09:27.405 { 00:09:27.405 "name": "BaseBdev4", 00:09:27.405 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:27.405 "is_configured": true, 00:09:27.405 "data_offset": 0, 00:09:27.405 "data_size": 65536 00:09:27.405 } 00:09:27.405 ] 00:09:27.405 }' 00:09:27.405 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.405 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.974 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.974 [2024-09-30 14:08:32.463902] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.975 "name": "Existed_Raid", 00:09:27.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.975 "strip_size_kb": 64, 00:09:27.975 "state": "configuring", 00:09:27.975 "raid_level": "raid0", 00:09:27.975 "superblock": false, 00:09:27.975 "num_base_bdevs": 4, 00:09:27.975 "num_base_bdevs_discovered": 3, 00:09:27.975 "num_base_bdevs_operational": 4, 00:09:27.975 "base_bdevs_list": [ 00:09:27.975 { 00:09:27.975 "name": "BaseBdev1", 00:09:27.975 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:27.975 "is_configured": true, 00:09:27.975 "data_offset": 0, 00:09:27.975 "data_size": 65536 00:09:27.975 }, 00:09:27.975 { 00:09:27.975 "name": null, 00:09:27.975 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:27.975 "is_configured": false, 00:09:27.975 "data_offset": 0, 00:09:27.975 "data_size": 65536 00:09:27.975 }, 00:09:27.975 { 00:09:27.975 "name": "BaseBdev3", 00:09:27.975 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:27.975 "is_configured": true, 00:09:27.975 "data_offset": 0, 00:09:27.975 "data_size": 65536 00:09:27.975 }, 00:09:27.975 { 00:09:27.975 "name": "BaseBdev4", 00:09:27.975 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:27.975 "is_configured": true, 00:09:27.975 "data_offset": 0, 00:09:27.975 "data_size": 65536 00:09:27.975 } 00:09:27.975 ] 00:09:27.975 }' 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.975 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.545 14:08:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.545 [2024-09-30 14:08:32.999021] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.545 "name": "Existed_Raid", 00:09:28.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.545 "strip_size_kb": 64, 00:09:28.545 "state": "configuring", 00:09:28.545 "raid_level": "raid0", 00:09:28.545 "superblock": false, 00:09:28.545 "num_base_bdevs": 4, 00:09:28.545 "num_base_bdevs_discovered": 2, 00:09:28.545 "num_base_bdevs_operational": 4, 00:09:28.545 "base_bdevs_list": [ 00:09:28.545 { 00:09:28.545 "name": null, 00:09:28.545 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:28.545 "is_configured": false, 00:09:28.545 "data_offset": 0, 00:09:28.545 "data_size": 65536 00:09:28.545 }, 00:09:28.545 { 00:09:28.545 "name": null, 00:09:28.545 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:28.545 "is_configured": false, 00:09:28.545 "data_offset": 0, 00:09:28.545 "data_size": 65536 00:09:28.545 }, 00:09:28.545 { 00:09:28.545 "name": "BaseBdev3", 00:09:28.545 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:28.545 "is_configured": true, 00:09:28.545 "data_offset": 0, 00:09:28.545 "data_size": 65536 00:09:28.545 }, 00:09:28.545 { 00:09:28.545 "name": "BaseBdev4", 00:09:28.545 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:28.545 "is_configured": true, 00:09:28.545 "data_offset": 0, 00:09:28.545 "data_size": 65536 00:09:28.545 } 00:09:28.545 ] 00:09:28.545 }' 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.545 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 [2024-09-30 14:08:33.520457] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.115 "name": "Existed_Raid", 00:09:29.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.115 "strip_size_kb": 64, 00:09:29.115 "state": "configuring", 00:09:29.115 "raid_level": "raid0", 00:09:29.115 "superblock": false, 00:09:29.115 "num_base_bdevs": 4, 00:09:29.115 "num_base_bdevs_discovered": 3, 00:09:29.115 "num_base_bdevs_operational": 4, 00:09:29.115 "base_bdevs_list": [ 00:09:29.115 { 00:09:29.115 "name": null, 00:09:29.115 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:29.115 "is_configured": false, 00:09:29.115 "data_offset": 0, 00:09:29.115 "data_size": 65536 00:09:29.115 }, 00:09:29.115 { 00:09:29.115 "name": "BaseBdev2", 00:09:29.115 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:29.115 "is_configured": true, 00:09:29.115 "data_offset": 0, 00:09:29.115 "data_size": 65536 00:09:29.115 }, 00:09:29.115 { 00:09:29.115 "name": "BaseBdev3", 00:09:29.115 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:29.115 "is_configured": true, 00:09:29.115 "data_offset": 0, 00:09:29.115 "data_size": 65536 00:09:29.115 }, 00:09:29.115 { 00:09:29.115 "name": "BaseBdev4", 00:09:29.115 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:29.115 "is_configured": true, 00:09:29.115 "data_offset": 0, 00:09:29.115 "data_size": 65536 00:09:29.115 } 00:09:29.115 ] 00:09:29.115 }' 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.115 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.375 14:08:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.375 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 46fab4c2-b38f-4890-9bec-55322e39e08e 00:09:29.375 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.375 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.375 [2024-09-30 14:08:34.030405] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:29.375 [2024-09-30 14:08:34.030529] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:29.375 [2024-09-30 14:08:34.030558] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:29.375 [2024-09-30 14:08:34.030836] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:09:29.635 [2024-09-30 14:08:34.030987] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:29.635 [2024-09-30 14:08:34.031023] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:29.635 [2024-09-30 14:08:34.031231] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:29.635 NewBaseBdev 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.635 [ 00:09:29.635 { 00:09:29.635 "name": "NewBaseBdev", 00:09:29.635 "aliases": [ 00:09:29.635 "46fab4c2-b38f-4890-9bec-55322e39e08e" 00:09:29.635 ], 00:09:29.635 "product_name": "Malloc disk", 00:09:29.635 "block_size": 512, 00:09:29.635 "num_blocks": 65536, 00:09:29.635 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:29.635 "assigned_rate_limits": { 00:09:29.635 "rw_ios_per_sec": 0, 00:09:29.635 "rw_mbytes_per_sec": 0, 00:09:29.635 "r_mbytes_per_sec": 0, 00:09:29.635 "w_mbytes_per_sec": 0 00:09:29.635 }, 00:09:29.635 "claimed": true, 00:09:29.635 "claim_type": "exclusive_write", 00:09:29.635 "zoned": false, 00:09:29.635 "supported_io_types": { 00:09:29.635 "read": true, 00:09:29.635 "write": true, 00:09:29.635 "unmap": true, 00:09:29.635 "flush": true, 00:09:29.635 "reset": true, 00:09:29.635 "nvme_admin": false, 00:09:29.635 "nvme_io": false, 00:09:29.635 "nvme_io_md": false, 00:09:29.635 "write_zeroes": true, 00:09:29.635 "zcopy": true, 00:09:29.635 "get_zone_info": false, 00:09:29.635 "zone_management": false, 00:09:29.635 "zone_append": false, 00:09:29.635 "compare": false, 00:09:29.635 "compare_and_write": false, 00:09:29.635 "abort": true, 00:09:29.635 "seek_hole": false, 00:09:29.635 "seek_data": false, 00:09:29.635 "copy": true, 00:09:29.635 "nvme_iov_md": false 00:09:29.635 }, 00:09:29.635 "memory_domains": [ 00:09:29.635 { 00:09:29.635 "dma_device_id": "system", 00:09:29.635 "dma_device_type": 1 00:09:29.635 }, 00:09:29.635 { 00:09:29.635 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.635 "dma_device_type": 2 00:09:29.635 } 00:09:29.635 ], 00:09:29.635 "driver_specific": {} 00:09:29.635 } 00:09:29.635 ] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.635 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.635 "name": "Existed_Raid", 00:09:29.635 "uuid": "b6012ad2-78c2-4054-b53f-16c1c1a446af", 00:09:29.635 "strip_size_kb": 64, 00:09:29.635 "state": "online", 00:09:29.635 "raid_level": "raid0", 00:09:29.635 "superblock": false, 00:09:29.635 "num_base_bdevs": 4, 00:09:29.635 "num_base_bdevs_discovered": 4, 00:09:29.635 "num_base_bdevs_operational": 4, 00:09:29.635 "base_bdevs_list": [ 00:09:29.635 { 00:09:29.635 "name": "NewBaseBdev", 00:09:29.635 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:29.635 "is_configured": true, 00:09:29.635 "data_offset": 0, 00:09:29.636 "data_size": 65536 00:09:29.636 }, 00:09:29.636 { 00:09:29.636 "name": "BaseBdev2", 00:09:29.636 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:29.636 "is_configured": true, 00:09:29.636 "data_offset": 0, 00:09:29.636 "data_size": 65536 00:09:29.636 }, 00:09:29.636 { 00:09:29.636 "name": "BaseBdev3", 00:09:29.636 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:29.636 "is_configured": true, 00:09:29.636 "data_offset": 0, 00:09:29.636 "data_size": 65536 00:09:29.636 }, 00:09:29.636 { 00:09:29.636 "name": "BaseBdev4", 00:09:29.636 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:29.636 "is_configured": true, 00:09:29.636 "data_offset": 0, 00:09:29.636 "data_size": 65536 00:09:29.636 } 00:09:29.636 ] 00:09:29.636 }' 00:09:29.636 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.636 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:29.915 [2024-09-30 14:08:34.501923] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:29.915 "name": "Existed_Raid", 00:09:29.915 "aliases": [ 00:09:29.915 "b6012ad2-78c2-4054-b53f-16c1c1a446af" 00:09:29.915 ], 00:09:29.915 "product_name": "Raid Volume", 00:09:29.915 "block_size": 512, 00:09:29.915 "num_blocks": 262144, 00:09:29.915 "uuid": "b6012ad2-78c2-4054-b53f-16c1c1a446af", 00:09:29.915 "assigned_rate_limits": { 00:09:29.915 "rw_ios_per_sec": 0, 00:09:29.915 "rw_mbytes_per_sec": 0, 00:09:29.915 "r_mbytes_per_sec": 0, 00:09:29.915 "w_mbytes_per_sec": 0 00:09:29.915 }, 00:09:29.915 "claimed": false, 00:09:29.915 "zoned": false, 00:09:29.915 "supported_io_types": { 00:09:29.915 "read": true, 00:09:29.915 "write": true, 00:09:29.915 "unmap": true, 00:09:29.915 "flush": true, 00:09:29.915 "reset": true, 00:09:29.915 "nvme_admin": false, 00:09:29.915 "nvme_io": false, 00:09:29.915 "nvme_io_md": false, 00:09:29.915 "write_zeroes": true, 00:09:29.915 "zcopy": false, 00:09:29.915 "get_zone_info": false, 00:09:29.915 "zone_management": false, 00:09:29.915 "zone_append": false, 00:09:29.915 "compare": false, 00:09:29.915 "compare_and_write": false, 00:09:29.915 "abort": false, 00:09:29.915 "seek_hole": false, 00:09:29.915 "seek_data": false, 00:09:29.915 "copy": false, 00:09:29.915 "nvme_iov_md": false 00:09:29.915 }, 00:09:29.915 "memory_domains": [ 00:09:29.915 { 00:09:29.915 "dma_device_id": "system", 00:09:29.915 "dma_device_type": 1 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.915 "dma_device_type": 2 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "system", 00:09:29.915 "dma_device_type": 1 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.915 "dma_device_type": 2 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "system", 00:09:29.915 "dma_device_type": 1 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.915 "dma_device_type": 2 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "system", 00:09:29.915 "dma_device_type": 1 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.915 "dma_device_type": 2 00:09:29.915 } 00:09:29.915 ], 00:09:29.915 "driver_specific": { 00:09:29.915 "raid": { 00:09:29.915 "uuid": "b6012ad2-78c2-4054-b53f-16c1c1a446af", 00:09:29.915 "strip_size_kb": 64, 00:09:29.915 "state": "online", 00:09:29.915 "raid_level": "raid0", 00:09:29.915 "superblock": false, 00:09:29.915 "num_base_bdevs": 4, 00:09:29.915 "num_base_bdevs_discovered": 4, 00:09:29.915 "num_base_bdevs_operational": 4, 00:09:29.915 "base_bdevs_list": [ 00:09:29.915 { 00:09:29.915 "name": "NewBaseBdev", 00:09:29.915 "uuid": "46fab4c2-b38f-4890-9bec-55322e39e08e", 00:09:29.915 "is_configured": true, 00:09:29.915 "data_offset": 0, 00:09:29.915 "data_size": 65536 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "name": "BaseBdev2", 00:09:29.915 "uuid": "38ddd4d6-5bfc-4a4b-b802-4037a8d3b25c", 00:09:29.915 "is_configured": true, 00:09:29.915 "data_offset": 0, 00:09:29.915 "data_size": 65536 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "name": "BaseBdev3", 00:09:29.915 "uuid": "fc9ac95a-c498-4ff6-9e6c-7ddffe5d5c91", 00:09:29.915 "is_configured": true, 00:09:29.915 "data_offset": 0, 00:09:29.915 "data_size": 65536 00:09:29.915 }, 00:09:29.915 { 00:09:29.915 "name": "BaseBdev4", 00:09:29.915 "uuid": "b6e0b777-9448-4934-8576-74e140ff430a", 00:09:29.915 "is_configured": true, 00:09:29.915 "data_offset": 0, 00:09:29.915 "data_size": 65536 00:09:29.915 } 00:09:29.915 ] 00:09:29.915 } 00:09:29.915 } 00:09:29.915 }' 00:09:29.915 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:30.191 BaseBdev2 00:09:30.191 BaseBdev3 00:09:30.191 BaseBdev4' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.191 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.192 [2024-09-30 14:08:34.833030] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:30.192 [2024-09-30 14:08:34.833100] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:30.192 [2024-09-30 14:08:34.833193] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:30.192 [2024-09-30 14:08:34.833273] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:30.192 [2024-09-30 14:08:34.833309] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81394 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 81394 ']' 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 81394 00:09:30.192 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81394 00:09:30.451 killing process with pid 81394 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81394' 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 81394 00:09:30.451 [2024-09-30 14:08:34.883513] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:30.451 14:08:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 81394 00:09:30.451 [2024-09-30 14:08:34.923578] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:30.710 14:08:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:30.710 00:09:30.710 real 0m9.511s 00:09:30.710 user 0m16.181s 00:09:30.710 sys 0m2.064s 00:09:30.710 14:08:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:30.710 ************************************ 00:09:30.710 END TEST raid_state_function_test 00:09:30.710 ************************************ 00:09:30.710 14:08:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.710 14:08:35 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:09:30.710 14:08:35 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:30.711 14:08:35 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:30.711 14:08:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:30.711 ************************************ 00:09:30.711 START TEST raid_state_function_test_sb 00:09:30.711 ************************************ 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 true 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82042 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82042' 00:09:30.711 Process raid pid: 82042 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82042 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 82042 ']' 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:30.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:30.711 14:08:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.711 [2024-09-30 14:08:35.346921] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:30.711 [2024-09-30 14:08:35.347050] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:30.970 [2024-09-30 14:08:35.478518] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:30.970 [2024-09-30 14:08:35.505061] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:30.970 [2024-09-30 14:08:35.549278] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:30.970 [2024-09-30 14:08:35.590419] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:30.970 [2024-09-30 14:08:35.590454] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.539 [2024-09-30 14:08:36.179168] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:31.539 [2024-09-30 14:08:36.179217] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:31.539 [2024-09-30 14:08:36.179236] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:31.539 [2024-09-30 14:08:36.179244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:31.539 [2024-09-30 14:08:36.179254] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:31.539 [2024-09-30 14:08:36.179260] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:31.539 [2024-09-30 14:08:36.179268] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:31.539 [2024-09-30 14:08:36.179274] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.539 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.799 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.799 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.799 "name": "Existed_Raid", 00:09:31.799 "uuid": "f0afb092-15cd-4ea9-810d-a3a9c4cc715a", 00:09:31.799 "strip_size_kb": 64, 00:09:31.799 "state": "configuring", 00:09:31.799 "raid_level": "raid0", 00:09:31.799 "superblock": true, 00:09:31.799 "num_base_bdevs": 4, 00:09:31.799 "num_base_bdevs_discovered": 0, 00:09:31.799 "num_base_bdevs_operational": 4, 00:09:31.799 "base_bdevs_list": [ 00:09:31.799 { 00:09:31.799 "name": "BaseBdev1", 00:09:31.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.799 "is_configured": false, 00:09:31.799 "data_offset": 0, 00:09:31.799 "data_size": 0 00:09:31.799 }, 00:09:31.799 { 00:09:31.799 "name": "BaseBdev2", 00:09:31.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.799 "is_configured": false, 00:09:31.799 "data_offset": 0, 00:09:31.799 "data_size": 0 00:09:31.799 }, 00:09:31.799 { 00:09:31.799 "name": "BaseBdev3", 00:09:31.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.799 "is_configured": false, 00:09:31.799 "data_offset": 0, 00:09:31.799 "data_size": 0 00:09:31.799 }, 00:09:31.799 { 00:09:31.799 "name": "BaseBdev4", 00:09:31.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.799 "is_configured": false, 00:09:31.799 "data_offset": 0, 00:09:31.799 "data_size": 0 00:09:31.799 } 00:09:31.799 ] 00:09:31.799 }' 00:09:31.799 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.799 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 [2024-09-30 14:08:36.614321] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:32.058 [2024-09-30 14:08:36.614397] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 [2024-09-30 14:08:36.622318] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:32.058 [2024-09-30 14:08:36.622391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:32.058 [2024-09-30 14:08:36.622419] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:32.058 [2024-09-30 14:08:36.622438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:32.058 [2024-09-30 14:08:36.622457] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:32.058 [2024-09-30 14:08:36.622475] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:32.058 [2024-09-30 14:08:36.622502] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:32.058 [2024-09-30 14:08:36.622520] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 [2024-09-30 14:08:36.638836] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:32.058 BaseBdev1 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.058 [ 00:09:32.058 { 00:09:32.058 "name": "BaseBdev1", 00:09:32.058 "aliases": [ 00:09:32.058 "e9a18aa6-fc25-47af-ba00-ec2bcc859dce" 00:09:32.058 ], 00:09:32.058 "product_name": "Malloc disk", 00:09:32.058 "block_size": 512, 00:09:32.058 "num_blocks": 65536, 00:09:32.058 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:32.058 "assigned_rate_limits": { 00:09:32.058 "rw_ios_per_sec": 0, 00:09:32.058 "rw_mbytes_per_sec": 0, 00:09:32.058 "r_mbytes_per_sec": 0, 00:09:32.058 "w_mbytes_per_sec": 0 00:09:32.058 }, 00:09:32.058 "claimed": true, 00:09:32.058 "claim_type": "exclusive_write", 00:09:32.058 "zoned": false, 00:09:32.058 "supported_io_types": { 00:09:32.058 "read": true, 00:09:32.058 "write": true, 00:09:32.058 "unmap": true, 00:09:32.058 "flush": true, 00:09:32.058 "reset": true, 00:09:32.058 "nvme_admin": false, 00:09:32.058 "nvme_io": false, 00:09:32.058 "nvme_io_md": false, 00:09:32.058 "write_zeroes": true, 00:09:32.058 "zcopy": true, 00:09:32.058 "get_zone_info": false, 00:09:32.058 "zone_management": false, 00:09:32.058 "zone_append": false, 00:09:32.058 "compare": false, 00:09:32.058 "compare_and_write": false, 00:09:32.058 "abort": true, 00:09:32.058 "seek_hole": false, 00:09:32.058 "seek_data": false, 00:09:32.058 "copy": true, 00:09:32.058 "nvme_iov_md": false 00:09:32.058 }, 00:09:32.058 "memory_domains": [ 00:09:32.058 { 00:09:32.058 "dma_device_id": "system", 00:09:32.058 "dma_device_type": 1 00:09:32.058 }, 00:09:32.058 { 00:09:32.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.058 "dma_device_type": 2 00:09:32.058 } 00:09:32.058 ], 00:09:32.058 "driver_specific": {} 00:09:32.058 } 00:09:32.058 ] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:32.058 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.059 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.318 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.318 "name": "Existed_Raid", 00:09:32.318 "uuid": "85f48abe-fd25-4d18-9226-047ce5b1f107", 00:09:32.318 "strip_size_kb": 64, 00:09:32.318 "state": "configuring", 00:09:32.318 "raid_level": "raid0", 00:09:32.318 "superblock": true, 00:09:32.318 "num_base_bdevs": 4, 00:09:32.318 "num_base_bdevs_discovered": 1, 00:09:32.318 "num_base_bdevs_operational": 4, 00:09:32.318 "base_bdevs_list": [ 00:09:32.318 { 00:09:32.318 "name": "BaseBdev1", 00:09:32.318 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:32.318 "is_configured": true, 00:09:32.318 "data_offset": 2048, 00:09:32.318 "data_size": 63488 00:09:32.318 }, 00:09:32.318 { 00:09:32.318 "name": "BaseBdev2", 00:09:32.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.318 "is_configured": false, 00:09:32.318 "data_offset": 0, 00:09:32.318 "data_size": 0 00:09:32.318 }, 00:09:32.318 { 00:09:32.318 "name": "BaseBdev3", 00:09:32.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.318 "is_configured": false, 00:09:32.318 "data_offset": 0, 00:09:32.318 "data_size": 0 00:09:32.318 }, 00:09:32.318 { 00:09:32.318 "name": "BaseBdev4", 00:09:32.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.318 "is_configured": false, 00:09:32.318 "data_offset": 0, 00:09:32.318 "data_size": 0 00:09:32.318 } 00:09:32.318 ] 00:09:32.318 }' 00:09:32.318 14:08:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.318 14:08:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.577 [2024-09-30 14:08:37.070160] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:32.577 [2024-09-30 14:08:37.070210] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.577 [2024-09-30 14:08:37.082159] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:32.577 [2024-09-30 14:08:37.084069] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:32.577 [2024-09-30 14:08:37.084138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:32.577 [2024-09-30 14:08:37.084167] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:32.577 [2024-09-30 14:08:37.084186] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:32.577 [2024-09-30 14:08:37.084204] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:32.577 [2024-09-30 14:08:37.084221] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.577 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.577 "name": "Existed_Raid", 00:09:32.577 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:32.577 "strip_size_kb": 64, 00:09:32.577 "state": "configuring", 00:09:32.577 "raid_level": "raid0", 00:09:32.577 "superblock": true, 00:09:32.577 "num_base_bdevs": 4, 00:09:32.577 "num_base_bdevs_discovered": 1, 00:09:32.577 "num_base_bdevs_operational": 4, 00:09:32.577 "base_bdevs_list": [ 00:09:32.577 { 00:09:32.577 "name": "BaseBdev1", 00:09:32.577 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:32.577 "is_configured": true, 00:09:32.577 "data_offset": 2048, 00:09:32.577 "data_size": 63488 00:09:32.577 }, 00:09:32.577 { 00:09:32.577 "name": "BaseBdev2", 00:09:32.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.577 "is_configured": false, 00:09:32.577 "data_offset": 0, 00:09:32.577 "data_size": 0 00:09:32.577 }, 00:09:32.577 { 00:09:32.577 "name": "BaseBdev3", 00:09:32.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.577 "is_configured": false, 00:09:32.577 "data_offset": 0, 00:09:32.577 "data_size": 0 00:09:32.577 }, 00:09:32.577 { 00:09:32.577 "name": "BaseBdev4", 00:09:32.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.578 "is_configured": false, 00:09:32.578 "data_offset": 0, 00:09:32.578 "data_size": 0 00:09:32.578 } 00:09:32.578 ] 00:09:32.578 }' 00:09:32.578 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.578 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.146 [2024-09-30 14:08:37.538826] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:33.146 BaseBdev2 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.146 [ 00:09:33.146 { 00:09:33.146 "name": "BaseBdev2", 00:09:33.146 "aliases": [ 00:09:33.146 "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8" 00:09:33.146 ], 00:09:33.146 "product_name": "Malloc disk", 00:09:33.146 "block_size": 512, 00:09:33.146 "num_blocks": 65536, 00:09:33.146 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:33.146 "assigned_rate_limits": { 00:09:33.146 "rw_ios_per_sec": 0, 00:09:33.146 "rw_mbytes_per_sec": 0, 00:09:33.146 "r_mbytes_per_sec": 0, 00:09:33.146 "w_mbytes_per_sec": 0 00:09:33.146 }, 00:09:33.146 "claimed": true, 00:09:33.146 "claim_type": "exclusive_write", 00:09:33.146 "zoned": false, 00:09:33.146 "supported_io_types": { 00:09:33.146 "read": true, 00:09:33.146 "write": true, 00:09:33.146 "unmap": true, 00:09:33.146 "flush": true, 00:09:33.146 "reset": true, 00:09:33.146 "nvme_admin": false, 00:09:33.146 "nvme_io": false, 00:09:33.146 "nvme_io_md": false, 00:09:33.146 "write_zeroes": true, 00:09:33.146 "zcopy": true, 00:09:33.146 "get_zone_info": false, 00:09:33.146 "zone_management": false, 00:09:33.146 "zone_append": false, 00:09:33.146 "compare": false, 00:09:33.146 "compare_and_write": false, 00:09:33.146 "abort": true, 00:09:33.146 "seek_hole": false, 00:09:33.146 "seek_data": false, 00:09:33.146 "copy": true, 00:09:33.146 "nvme_iov_md": false 00:09:33.146 }, 00:09:33.146 "memory_domains": [ 00:09:33.146 { 00:09:33.146 "dma_device_id": "system", 00:09:33.146 "dma_device_type": 1 00:09:33.146 }, 00:09:33.146 { 00:09:33.146 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.146 "dma_device_type": 2 00:09:33.146 } 00:09:33.146 ], 00:09:33.146 "driver_specific": {} 00:09:33.146 } 00:09:33.146 ] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.146 "name": "Existed_Raid", 00:09:33.146 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:33.146 "strip_size_kb": 64, 00:09:33.146 "state": "configuring", 00:09:33.146 "raid_level": "raid0", 00:09:33.146 "superblock": true, 00:09:33.146 "num_base_bdevs": 4, 00:09:33.146 "num_base_bdevs_discovered": 2, 00:09:33.146 "num_base_bdevs_operational": 4, 00:09:33.146 "base_bdevs_list": [ 00:09:33.146 { 00:09:33.146 "name": "BaseBdev1", 00:09:33.146 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:33.146 "is_configured": true, 00:09:33.146 "data_offset": 2048, 00:09:33.146 "data_size": 63488 00:09:33.146 }, 00:09:33.146 { 00:09:33.146 "name": "BaseBdev2", 00:09:33.146 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:33.146 "is_configured": true, 00:09:33.146 "data_offset": 2048, 00:09:33.146 "data_size": 63488 00:09:33.146 }, 00:09:33.146 { 00:09:33.146 "name": "BaseBdev3", 00:09:33.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.146 "is_configured": false, 00:09:33.146 "data_offset": 0, 00:09:33.146 "data_size": 0 00:09:33.146 }, 00:09:33.146 { 00:09:33.146 "name": "BaseBdev4", 00:09:33.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.146 "is_configured": false, 00:09:33.146 "data_offset": 0, 00:09:33.146 "data_size": 0 00:09:33.146 } 00:09:33.146 ] 00:09:33.146 }' 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.146 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.405 [2024-09-30 14:08:37.988920] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:33.405 BaseBdev3 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.405 14:08:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.405 [ 00:09:33.405 { 00:09:33.405 "name": "BaseBdev3", 00:09:33.405 "aliases": [ 00:09:33.405 "abd3b657-7453-4051-833f-d6a81d664ca3" 00:09:33.405 ], 00:09:33.405 "product_name": "Malloc disk", 00:09:33.405 "block_size": 512, 00:09:33.405 "num_blocks": 65536, 00:09:33.405 "uuid": "abd3b657-7453-4051-833f-d6a81d664ca3", 00:09:33.405 "assigned_rate_limits": { 00:09:33.405 "rw_ios_per_sec": 0, 00:09:33.405 "rw_mbytes_per_sec": 0, 00:09:33.405 "r_mbytes_per_sec": 0, 00:09:33.405 "w_mbytes_per_sec": 0 00:09:33.405 }, 00:09:33.405 "claimed": true, 00:09:33.405 "claim_type": "exclusive_write", 00:09:33.405 "zoned": false, 00:09:33.405 "supported_io_types": { 00:09:33.405 "read": true, 00:09:33.405 "write": true, 00:09:33.405 "unmap": true, 00:09:33.405 "flush": true, 00:09:33.405 "reset": true, 00:09:33.405 "nvme_admin": false, 00:09:33.405 "nvme_io": false, 00:09:33.405 "nvme_io_md": false, 00:09:33.405 "write_zeroes": true, 00:09:33.405 "zcopy": true, 00:09:33.405 "get_zone_info": false, 00:09:33.405 "zone_management": false, 00:09:33.405 "zone_append": false, 00:09:33.405 "compare": false, 00:09:33.405 "compare_and_write": false, 00:09:33.405 "abort": true, 00:09:33.405 "seek_hole": false, 00:09:33.405 "seek_data": false, 00:09:33.405 "copy": true, 00:09:33.405 "nvme_iov_md": false 00:09:33.405 }, 00:09:33.405 "memory_domains": [ 00:09:33.405 { 00:09:33.405 "dma_device_id": "system", 00:09:33.405 "dma_device_type": 1 00:09:33.405 }, 00:09:33.405 { 00:09:33.405 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.405 "dma_device_type": 2 00:09:33.405 } 00:09:33.405 ], 00:09:33.405 "driver_specific": {} 00:09:33.405 } 00:09:33.405 ] 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.405 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.406 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.665 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.665 "name": "Existed_Raid", 00:09:33.665 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:33.665 "strip_size_kb": 64, 00:09:33.665 "state": "configuring", 00:09:33.665 "raid_level": "raid0", 00:09:33.665 "superblock": true, 00:09:33.665 "num_base_bdevs": 4, 00:09:33.665 "num_base_bdevs_discovered": 3, 00:09:33.665 "num_base_bdevs_operational": 4, 00:09:33.665 "base_bdevs_list": [ 00:09:33.665 { 00:09:33.665 "name": "BaseBdev1", 00:09:33.665 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:33.665 "is_configured": true, 00:09:33.665 "data_offset": 2048, 00:09:33.665 "data_size": 63488 00:09:33.665 }, 00:09:33.665 { 00:09:33.665 "name": "BaseBdev2", 00:09:33.665 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:33.665 "is_configured": true, 00:09:33.665 "data_offset": 2048, 00:09:33.665 "data_size": 63488 00:09:33.665 }, 00:09:33.665 { 00:09:33.665 "name": "BaseBdev3", 00:09:33.665 "uuid": "abd3b657-7453-4051-833f-d6a81d664ca3", 00:09:33.665 "is_configured": true, 00:09:33.665 "data_offset": 2048, 00:09:33.665 "data_size": 63488 00:09:33.665 }, 00:09:33.665 { 00:09:33.665 "name": "BaseBdev4", 00:09:33.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.665 "is_configured": false, 00:09:33.665 "data_offset": 0, 00:09:33.665 "data_size": 0 00:09:33.665 } 00:09:33.665 ] 00:09:33.665 }' 00:09:33.665 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.665 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.925 [2024-09-30 14:08:38.435074] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:33.925 BaseBdev4 00:09:33.925 [2024-09-30 14:08:38.435346] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:33.925 [2024-09-30 14:08:38.435367] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:33.925 [2024-09-30 14:08:38.435676] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:33.925 [2024-09-30 14:08:38.435801] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:33.925 [2024-09-30 14:08:38.435812] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:33.925 [2024-09-30 14:08:38.435946] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.925 [ 00:09:33.925 { 00:09:33.925 "name": "BaseBdev4", 00:09:33.925 "aliases": [ 00:09:33.925 "3f8af0ec-9951-40ae-a8c9-6f05f9e604f7" 00:09:33.925 ], 00:09:33.925 "product_name": "Malloc disk", 00:09:33.925 "block_size": 512, 00:09:33.925 "num_blocks": 65536, 00:09:33.925 "uuid": "3f8af0ec-9951-40ae-a8c9-6f05f9e604f7", 00:09:33.925 "assigned_rate_limits": { 00:09:33.925 "rw_ios_per_sec": 0, 00:09:33.925 "rw_mbytes_per_sec": 0, 00:09:33.925 "r_mbytes_per_sec": 0, 00:09:33.925 "w_mbytes_per_sec": 0 00:09:33.925 }, 00:09:33.925 "claimed": true, 00:09:33.925 "claim_type": "exclusive_write", 00:09:33.925 "zoned": false, 00:09:33.925 "supported_io_types": { 00:09:33.925 "read": true, 00:09:33.925 "write": true, 00:09:33.925 "unmap": true, 00:09:33.925 "flush": true, 00:09:33.925 "reset": true, 00:09:33.925 "nvme_admin": false, 00:09:33.925 "nvme_io": false, 00:09:33.925 "nvme_io_md": false, 00:09:33.925 "write_zeroes": true, 00:09:33.925 "zcopy": true, 00:09:33.925 "get_zone_info": false, 00:09:33.925 "zone_management": false, 00:09:33.925 "zone_append": false, 00:09:33.925 "compare": false, 00:09:33.925 "compare_and_write": false, 00:09:33.925 "abort": true, 00:09:33.925 "seek_hole": false, 00:09:33.925 "seek_data": false, 00:09:33.925 "copy": true, 00:09:33.925 "nvme_iov_md": false 00:09:33.925 }, 00:09:33.925 "memory_domains": [ 00:09:33.925 { 00:09:33.925 "dma_device_id": "system", 00:09:33.925 "dma_device_type": 1 00:09:33.925 }, 00:09:33.925 { 00:09:33.925 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.925 "dma_device_type": 2 00:09:33.925 } 00:09:33.925 ], 00:09:33.925 "driver_specific": {} 00:09:33.925 } 00:09:33.925 ] 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.925 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.926 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.926 "name": "Existed_Raid", 00:09:33.926 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:33.926 "strip_size_kb": 64, 00:09:33.926 "state": "online", 00:09:33.926 "raid_level": "raid0", 00:09:33.926 "superblock": true, 00:09:33.926 "num_base_bdevs": 4, 00:09:33.926 "num_base_bdevs_discovered": 4, 00:09:33.926 "num_base_bdevs_operational": 4, 00:09:33.926 "base_bdevs_list": [ 00:09:33.926 { 00:09:33.926 "name": "BaseBdev1", 00:09:33.926 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:33.926 "is_configured": true, 00:09:33.926 "data_offset": 2048, 00:09:33.926 "data_size": 63488 00:09:33.926 }, 00:09:33.926 { 00:09:33.926 "name": "BaseBdev2", 00:09:33.926 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:33.926 "is_configured": true, 00:09:33.926 "data_offset": 2048, 00:09:33.926 "data_size": 63488 00:09:33.926 }, 00:09:33.926 { 00:09:33.926 "name": "BaseBdev3", 00:09:33.926 "uuid": "abd3b657-7453-4051-833f-d6a81d664ca3", 00:09:33.926 "is_configured": true, 00:09:33.926 "data_offset": 2048, 00:09:33.926 "data_size": 63488 00:09:33.926 }, 00:09:33.926 { 00:09:33.926 "name": "BaseBdev4", 00:09:33.926 "uuid": "3f8af0ec-9951-40ae-a8c9-6f05f9e604f7", 00:09:33.926 "is_configured": true, 00:09:33.926 "data_offset": 2048, 00:09:33.926 "data_size": 63488 00:09:33.926 } 00:09:33.926 ] 00:09:33.926 }' 00:09:33.926 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.926 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.500 [2024-09-30 14:08:38.946572] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.500 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:34.500 "name": "Existed_Raid", 00:09:34.500 "aliases": [ 00:09:34.500 "65f4661c-c1ee-4b94-9073-7600b9d9bd6a" 00:09:34.500 ], 00:09:34.500 "product_name": "Raid Volume", 00:09:34.500 "block_size": 512, 00:09:34.500 "num_blocks": 253952, 00:09:34.500 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:34.500 "assigned_rate_limits": { 00:09:34.500 "rw_ios_per_sec": 0, 00:09:34.500 "rw_mbytes_per_sec": 0, 00:09:34.500 "r_mbytes_per_sec": 0, 00:09:34.500 "w_mbytes_per_sec": 0 00:09:34.500 }, 00:09:34.500 "claimed": false, 00:09:34.500 "zoned": false, 00:09:34.500 "supported_io_types": { 00:09:34.500 "read": true, 00:09:34.500 "write": true, 00:09:34.500 "unmap": true, 00:09:34.500 "flush": true, 00:09:34.500 "reset": true, 00:09:34.500 "nvme_admin": false, 00:09:34.500 "nvme_io": false, 00:09:34.500 "nvme_io_md": false, 00:09:34.500 "write_zeroes": true, 00:09:34.500 "zcopy": false, 00:09:34.500 "get_zone_info": false, 00:09:34.500 "zone_management": false, 00:09:34.500 "zone_append": false, 00:09:34.500 "compare": false, 00:09:34.500 "compare_and_write": false, 00:09:34.500 "abort": false, 00:09:34.500 "seek_hole": false, 00:09:34.500 "seek_data": false, 00:09:34.500 "copy": false, 00:09:34.500 "nvme_iov_md": false 00:09:34.500 }, 00:09:34.500 "memory_domains": [ 00:09:34.500 { 00:09:34.500 "dma_device_id": "system", 00:09:34.500 "dma_device_type": 1 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.500 "dma_device_type": 2 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "system", 00:09:34.500 "dma_device_type": 1 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.500 "dma_device_type": 2 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "system", 00:09:34.500 "dma_device_type": 1 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.500 "dma_device_type": 2 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "system", 00:09:34.500 "dma_device_type": 1 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.500 "dma_device_type": 2 00:09:34.500 } 00:09:34.500 ], 00:09:34.500 "driver_specific": { 00:09:34.500 "raid": { 00:09:34.500 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:34.500 "strip_size_kb": 64, 00:09:34.500 "state": "online", 00:09:34.500 "raid_level": "raid0", 00:09:34.500 "superblock": true, 00:09:34.500 "num_base_bdevs": 4, 00:09:34.500 "num_base_bdevs_discovered": 4, 00:09:34.500 "num_base_bdevs_operational": 4, 00:09:34.500 "base_bdevs_list": [ 00:09:34.500 { 00:09:34.500 "name": "BaseBdev1", 00:09:34.500 "uuid": "e9a18aa6-fc25-47af-ba00-ec2bcc859dce", 00:09:34.500 "is_configured": true, 00:09:34.500 "data_offset": 2048, 00:09:34.500 "data_size": 63488 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "name": "BaseBdev2", 00:09:34.500 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:34.500 "is_configured": true, 00:09:34.500 "data_offset": 2048, 00:09:34.500 "data_size": 63488 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "name": "BaseBdev3", 00:09:34.500 "uuid": "abd3b657-7453-4051-833f-d6a81d664ca3", 00:09:34.500 "is_configured": true, 00:09:34.500 "data_offset": 2048, 00:09:34.500 "data_size": 63488 00:09:34.500 }, 00:09:34.500 { 00:09:34.500 "name": "BaseBdev4", 00:09:34.500 "uuid": "3f8af0ec-9951-40ae-a8c9-6f05f9e604f7", 00:09:34.500 "is_configured": true, 00:09:34.500 "data_offset": 2048, 00:09:34.500 "data_size": 63488 00:09:34.500 } 00:09:34.500 ] 00:09:34.500 } 00:09:34.500 } 00:09:34.500 }' 00:09:34.501 14:08:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:34.501 BaseBdev2 00:09:34.501 BaseBdev3 00:09:34.501 BaseBdev4' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.501 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.760 [2024-09-30 14:08:39.253781] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:34.760 [2024-09-30 14:08:39.253843] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:34.760 [2024-09-30 14:08:39.253923] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.760 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.761 "name": "Existed_Raid", 00:09:34.761 "uuid": "65f4661c-c1ee-4b94-9073-7600b9d9bd6a", 00:09:34.761 "strip_size_kb": 64, 00:09:34.761 "state": "offline", 00:09:34.761 "raid_level": "raid0", 00:09:34.761 "superblock": true, 00:09:34.761 "num_base_bdevs": 4, 00:09:34.761 "num_base_bdevs_discovered": 3, 00:09:34.761 "num_base_bdevs_operational": 3, 00:09:34.761 "base_bdevs_list": [ 00:09:34.761 { 00:09:34.761 "name": null, 00:09:34.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.761 "is_configured": false, 00:09:34.761 "data_offset": 0, 00:09:34.761 "data_size": 63488 00:09:34.761 }, 00:09:34.761 { 00:09:34.761 "name": "BaseBdev2", 00:09:34.761 "uuid": "a70f5c6b-a6bd-48e8-b3cc-d90db86a1dd8", 00:09:34.761 "is_configured": true, 00:09:34.761 "data_offset": 2048, 00:09:34.761 "data_size": 63488 00:09:34.761 }, 00:09:34.761 { 00:09:34.761 "name": "BaseBdev3", 00:09:34.761 "uuid": "abd3b657-7453-4051-833f-d6a81d664ca3", 00:09:34.761 "is_configured": true, 00:09:34.761 "data_offset": 2048, 00:09:34.761 "data_size": 63488 00:09:34.761 }, 00:09:34.761 { 00:09:34.761 "name": "BaseBdev4", 00:09:34.761 "uuid": "3f8af0ec-9951-40ae-a8c9-6f05f9e604f7", 00:09:34.761 "is_configured": true, 00:09:34.761 "data_offset": 2048, 00:09:34.761 "data_size": 63488 00:09:34.761 } 00:09:34.761 ] 00:09:34.761 }' 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.761 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:35.020 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.279 [2024-09-30 14:08:39.716340] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.279 [2024-09-30 14:08:39.767289] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:35.279 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.280 [2024-09-30 14:08:39.838051] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:35.280 [2024-09-30 14:08:39.838105] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.280 BaseBdev2 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.280 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.280 [ 00:09:35.280 { 00:09:35.280 "name": "BaseBdev2", 00:09:35.280 "aliases": [ 00:09:35.280 "abf4864d-b0b9-41e5-8873-4c024493d36f" 00:09:35.280 ], 00:09:35.280 "product_name": "Malloc disk", 00:09:35.280 "block_size": 512, 00:09:35.280 "num_blocks": 65536, 00:09:35.280 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:35.280 "assigned_rate_limits": { 00:09:35.280 "rw_ios_per_sec": 0, 00:09:35.280 "rw_mbytes_per_sec": 0, 00:09:35.280 "r_mbytes_per_sec": 0, 00:09:35.280 "w_mbytes_per_sec": 0 00:09:35.280 }, 00:09:35.280 "claimed": false, 00:09:35.280 "zoned": false, 00:09:35.280 "supported_io_types": { 00:09:35.280 "read": true, 00:09:35.280 "write": true, 00:09:35.280 "unmap": true, 00:09:35.280 "flush": true, 00:09:35.280 "reset": true, 00:09:35.280 "nvme_admin": false, 00:09:35.280 "nvme_io": false, 00:09:35.280 "nvme_io_md": false, 00:09:35.280 "write_zeroes": true, 00:09:35.280 "zcopy": true, 00:09:35.280 "get_zone_info": false, 00:09:35.280 "zone_management": false, 00:09:35.280 "zone_append": false, 00:09:35.280 "compare": false, 00:09:35.280 "compare_and_write": false, 00:09:35.280 "abort": true, 00:09:35.280 "seek_hole": false, 00:09:35.280 "seek_data": false, 00:09:35.280 "copy": true, 00:09:35.280 "nvme_iov_md": false 00:09:35.280 }, 00:09:35.280 "memory_domains": [ 00:09:35.280 { 00:09:35.545 "dma_device_id": "system", 00:09:35.545 "dma_device_type": 1 00:09:35.545 }, 00:09:35.545 { 00:09:35.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.545 "dma_device_type": 2 00:09:35.545 } 00:09:35.545 ], 00:09:35.545 "driver_specific": {} 00:09:35.545 } 00:09:35.545 ] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 BaseBdev3 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 [ 00:09:35.545 { 00:09:35.545 "name": "BaseBdev3", 00:09:35.545 "aliases": [ 00:09:35.545 "2472ea7b-8bc6-483b-ba04-f17f0596ca76" 00:09:35.545 ], 00:09:35.545 "product_name": "Malloc disk", 00:09:35.545 "block_size": 512, 00:09:35.545 "num_blocks": 65536, 00:09:35.545 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:35.545 "assigned_rate_limits": { 00:09:35.545 "rw_ios_per_sec": 0, 00:09:35.545 "rw_mbytes_per_sec": 0, 00:09:35.545 "r_mbytes_per_sec": 0, 00:09:35.545 "w_mbytes_per_sec": 0 00:09:35.545 }, 00:09:35.545 "claimed": false, 00:09:35.545 "zoned": false, 00:09:35.545 "supported_io_types": { 00:09:35.545 "read": true, 00:09:35.545 "write": true, 00:09:35.545 "unmap": true, 00:09:35.545 "flush": true, 00:09:35.545 "reset": true, 00:09:35.545 "nvme_admin": false, 00:09:35.545 "nvme_io": false, 00:09:35.545 "nvme_io_md": false, 00:09:35.545 "write_zeroes": true, 00:09:35.545 "zcopy": true, 00:09:35.545 "get_zone_info": false, 00:09:35.545 "zone_management": false, 00:09:35.545 "zone_append": false, 00:09:35.545 "compare": false, 00:09:35.545 "compare_and_write": false, 00:09:35.545 "abort": true, 00:09:35.545 "seek_hole": false, 00:09:35.545 "seek_data": false, 00:09:35.545 "copy": true, 00:09:35.545 "nvme_iov_md": false 00:09:35.545 }, 00:09:35.545 "memory_domains": [ 00:09:35.545 { 00:09:35.545 "dma_device_id": "system", 00:09:35.545 "dma_device_type": 1 00:09:35.545 }, 00:09:35.545 { 00:09:35.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.545 "dma_device_type": 2 00:09:35.545 } 00:09:35.545 ], 00:09:35.545 "driver_specific": {} 00:09:35.545 } 00:09:35.545 ] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 BaseBdev4 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.545 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.545 [ 00:09:35.545 { 00:09:35.545 "name": "BaseBdev4", 00:09:35.545 "aliases": [ 00:09:35.545 "47489534-458c-498e-9b43-a4dea536b725" 00:09:35.545 ], 00:09:35.545 "product_name": "Malloc disk", 00:09:35.545 "block_size": 512, 00:09:35.545 "num_blocks": 65536, 00:09:35.545 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:35.545 "assigned_rate_limits": { 00:09:35.545 "rw_ios_per_sec": 0, 00:09:35.545 "rw_mbytes_per_sec": 0, 00:09:35.545 "r_mbytes_per_sec": 0, 00:09:35.545 "w_mbytes_per_sec": 0 00:09:35.545 }, 00:09:35.545 "claimed": false, 00:09:35.545 "zoned": false, 00:09:35.546 "supported_io_types": { 00:09:35.546 "read": true, 00:09:35.546 "write": true, 00:09:35.546 "unmap": true, 00:09:35.546 "flush": true, 00:09:35.546 "reset": true, 00:09:35.546 "nvme_admin": false, 00:09:35.546 "nvme_io": false, 00:09:35.546 "nvme_io_md": false, 00:09:35.546 "write_zeroes": true, 00:09:35.546 "zcopy": true, 00:09:35.546 "get_zone_info": false, 00:09:35.546 "zone_management": false, 00:09:35.546 "zone_append": false, 00:09:35.546 "compare": false, 00:09:35.546 "compare_and_write": false, 00:09:35.546 "abort": true, 00:09:35.546 "seek_hole": false, 00:09:35.546 "seek_data": false, 00:09:35.546 "copy": true, 00:09:35.546 "nvme_iov_md": false 00:09:35.546 }, 00:09:35.546 "memory_domains": [ 00:09:35.546 { 00:09:35.546 "dma_device_id": "system", 00:09:35.546 "dma_device_type": 1 00:09:35.546 }, 00:09:35.546 { 00:09:35.546 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.546 "dma_device_type": 2 00:09:35.546 } 00:09:35.546 ], 00:09:35.546 "driver_specific": {} 00:09:35.546 } 00:09:35.546 ] 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.546 [2024-09-30 14:08:40.053145] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:35.546 [2024-09-30 14:08:40.053229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:35.546 [2024-09-30 14:08:40.053266] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:35.546 [2024-09-30 14:08:40.055054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:35.546 [2024-09-30 14:08:40.055135] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.546 "name": "Existed_Raid", 00:09:35.546 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:35.546 "strip_size_kb": 64, 00:09:35.546 "state": "configuring", 00:09:35.546 "raid_level": "raid0", 00:09:35.546 "superblock": true, 00:09:35.546 "num_base_bdevs": 4, 00:09:35.546 "num_base_bdevs_discovered": 3, 00:09:35.546 "num_base_bdevs_operational": 4, 00:09:35.546 "base_bdevs_list": [ 00:09:35.546 { 00:09:35.546 "name": "BaseBdev1", 00:09:35.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.546 "is_configured": false, 00:09:35.546 "data_offset": 0, 00:09:35.546 "data_size": 0 00:09:35.546 }, 00:09:35.546 { 00:09:35.546 "name": "BaseBdev2", 00:09:35.546 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:35.546 "is_configured": true, 00:09:35.546 "data_offset": 2048, 00:09:35.546 "data_size": 63488 00:09:35.546 }, 00:09:35.546 { 00:09:35.546 "name": "BaseBdev3", 00:09:35.546 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:35.546 "is_configured": true, 00:09:35.546 "data_offset": 2048, 00:09:35.546 "data_size": 63488 00:09:35.546 }, 00:09:35.546 { 00:09:35.546 "name": "BaseBdev4", 00:09:35.546 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:35.546 "is_configured": true, 00:09:35.546 "data_offset": 2048, 00:09:35.546 "data_size": 63488 00:09:35.546 } 00:09:35.546 ] 00:09:35.546 }' 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.546 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.115 [2024-09-30 14:08:40.480395] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.115 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.116 "name": "Existed_Raid", 00:09:36.116 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:36.116 "strip_size_kb": 64, 00:09:36.116 "state": "configuring", 00:09:36.116 "raid_level": "raid0", 00:09:36.116 "superblock": true, 00:09:36.116 "num_base_bdevs": 4, 00:09:36.116 "num_base_bdevs_discovered": 2, 00:09:36.116 "num_base_bdevs_operational": 4, 00:09:36.116 "base_bdevs_list": [ 00:09:36.116 { 00:09:36.116 "name": "BaseBdev1", 00:09:36.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.116 "is_configured": false, 00:09:36.116 "data_offset": 0, 00:09:36.116 "data_size": 0 00:09:36.116 }, 00:09:36.116 { 00:09:36.116 "name": null, 00:09:36.116 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:36.116 "is_configured": false, 00:09:36.116 "data_offset": 0, 00:09:36.116 "data_size": 63488 00:09:36.116 }, 00:09:36.116 { 00:09:36.116 "name": "BaseBdev3", 00:09:36.116 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:36.116 "is_configured": true, 00:09:36.116 "data_offset": 2048, 00:09:36.116 "data_size": 63488 00:09:36.116 }, 00:09:36.116 { 00:09:36.116 "name": "BaseBdev4", 00:09:36.116 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:36.116 "is_configured": true, 00:09:36.116 "data_offset": 2048, 00:09:36.116 "data_size": 63488 00:09:36.116 } 00:09:36.116 ] 00:09:36.116 }' 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.116 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.376 [2024-09-30 14:08:40.974319] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:36.376 BaseBdev1 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.376 14:08:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.376 [ 00:09:36.376 { 00:09:36.376 "name": "BaseBdev1", 00:09:36.376 "aliases": [ 00:09:36.376 "8e45d35a-6f33-40a1-93ba-342d91b2f552" 00:09:36.376 ], 00:09:36.376 "product_name": "Malloc disk", 00:09:36.376 "block_size": 512, 00:09:36.376 "num_blocks": 65536, 00:09:36.376 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:36.376 "assigned_rate_limits": { 00:09:36.376 "rw_ios_per_sec": 0, 00:09:36.376 "rw_mbytes_per_sec": 0, 00:09:36.376 "r_mbytes_per_sec": 0, 00:09:36.376 "w_mbytes_per_sec": 0 00:09:36.376 }, 00:09:36.376 "claimed": true, 00:09:36.376 "claim_type": "exclusive_write", 00:09:36.376 "zoned": false, 00:09:36.376 "supported_io_types": { 00:09:36.376 "read": true, 00:09:36.376 "write": true, 00:09:36.376 "unmap": true, 00:09:36.376 "flush": true, 00:09:36.376 "reset": true, 00:09:36.376 "nvme_admin": false, 00:09:36.376 "nvme_io": false, 00:09:36.376 "nvme_io_md": false, 00:09:36.376 "write_zeroes": true, 00:09:36.376 "zcopy": true, 00:09:36.376 "get_zone_info": false, 00:09:36.376 "zone_management": false, 00:09:36.376 "zone_append": false, 00:09:36.376 "compare": false, 00:09:36.376 "compare_and_write": false, 00:09:36.376 "abort": true, 00:09:36.376 "seek_hole": false, 00:09:36.376 "seek_data": false, 00:09:36.376 "copy": true, 00:09:36.376 "nvme_iov_md": false 00:09:36.376 }, 00:09:36.376 "memory_domains": [ 00:09:36.376 { 00:09:36.376 "dma_device_id": "system", 00:09:36.376 "dma_device_type": 1 00:09:36.377 }, 00:09:36.377 { 00:09:36.377 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.377 "dma_device_type": 2 00:09:36.377 } 00:09:36.377 ], 00:09:36.377 "driver_specific": {} 00:09:36.377 } 00:09:36.377 ] 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.377 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.636 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.636 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.636 "name": "Existed_Raid", 00:09:36.636 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:36.636 "strip_size_kb": 64, 00:09:36.636 "state": "configuring", 00:09:36.637 "raid_level": "raid0", 00:09:36.637 "superblock": true, 00:09:36.637 "num_base_bdevs": 4, 00:09:36.637 "num_base_bdevs_discovered": 3, 00:09:36.637 "num_base_bdevs_operational": 4, 00:09:36.637 "base_bdevs_list": [ 00:09:36.637 { 00:09:36.637 "name": "BaseBdev1", 00:09:36.637 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:36.637 "is_configured": true, 00:09:36.637 "data_offset": 2048, 00:09:36.637 "data_size": 63488 00:09:36.637 }, 00:09:36.637 { 00:09:36.637 "name": null, 00:09:36.637 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:36.637 "is_configured": false, 00:09:36.637 "data_offset": 0, 00:09:36.637 "data_size": 63488 00:09:36.637 }, 00:09:36.637 { 00:09:36.637 "name": "BaseBdev3", 00:09:36.637 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:36.637 "is_configured": true, 00:09:36.637 "data_offset": 2048, 00:09:36.637 "data_size": 63488 00:09:36.637 }, 00:09:36.637 { 00:09:36.637 "name": "BaseBdev4", 00:09:36.637 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:36.637 "is_configured": true, 00:09:36.637 "data_offset": 2048, 00:09:36.637 "data_size": 63488 00:09:36.637 } 00:09:36.637 ] 00:09:36.637 }' 00:09:36.637 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.637 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.896 [2024-09-30 14:08:41.513447] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.896 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.897 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.156 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.156 "name": "Existed_Raid", 00:09:37.156 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:37.156 "strip_size_kb": 64, 00:09:37.156 "state": "configuring", 00:09:37.156 "raid_level": "raid0", 00:09:37.156 "superblock": true, 00:09:37.156 "num_base_bdevs": 4, 00:09:37.156 "num_base_bdevs_discovered": 2, 00:09:37.156 "num_base_bdevs_operational": 4, 00:09:37.156 "base_bdevs_list": [ 00:09:37.156 { 00:09:37.156 "name": "BaseBdev1", 00:09:37.156 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:37.156 "is_configured": true, 00:09:37.156 "data_offset": 2048, 00:09:37.156 "data_size": 63488 00:09:37.156 }, 00:09:37.156 { 00:09:37.156 "name": null, 00:09:37.156 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:37.156 "is_configured": false, 00:09:37.156 "data_offset": 0, 00:09:37.156 "data_size": 63488 00:09:37.156 }, 00:09:37.156 { 00:09:37.156 "name": null, 00:09:37.156 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:37.156 "is_configured": false, 00:09:37.156 "data_offset": 0, 00:09:37.156 "data_size": 63488 00:09:37.156 }, 00:09:37.156 { 00:09:37.156 "name": "BaseBdev4", 00:09:37.156 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:37.156 "is_configured": true, 00:09:37.156 "data_offset": 2048, 00:09:37.156 "data_size": 63488 00:09:37.156 } 00:09:37.156 ] 00:09:37.156 }' 00:09:37.156 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.156 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.416 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.416 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.416 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.416 14:08:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:37.416 14:08:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.416 [2024-09-30 14:08:42.032600] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.416 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.675 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.675 "name": "Existed_Raid", 00:09:37.675 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:37.675 "strip_size_kb": 64, 00:09:37.675 "state": "configuring", 00:09:37.675 "raid_level": "raid0", 00:09:37.675 "superblock": true, 00:09:37.675 "num_base_bdevs": 4, 00:09:37.675 "num_base_bdevs_discovered": 3, 00:09:37.675 "num_base_bdevs_operational": 4, 00:09:37.675 "base_bdevs_list": [ 00:09:37.675 { 00:09:37.675 "name": "BaseBdev1", 00:09:37.675 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:37.675 "is_configured": true, 00:09:37.675 "data_offset": 2048, 00:09:37.675 "data_size": 63488 00:09:37.675 }, 00:09:37.675 { 00:09:37.675 "name": null, 00:09:37.675 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:37.675 "is_configured": false, 00:09:37.675 "data_offset": 0, 00:09:37.675 "data_size": 63488 00:09:37.675 }, 00:09:37.675 { 00:09:37.675 "name": "BaseBdev3", 00:09:37.675 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:37.675 "is_configured": true, 00:09:37.675 "data_offset": 2048, 00:09:37.675 "data_size": 63488 00:09:37.675 }, 00:09:37.675 { 00:09:37.675 "name": "BaseBdev4", 00:09:37.675 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:37.675 "is_configured": true, 00:09:37.675 "data_offset": 2048, 00:09:37.675 "data_size": 63488 00:09:37.675 } 00:09:37.675 ] 00:09:37.675 }' 00:09:37.675 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.675 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.935 [2024-09-30 14:08:42.515773] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.935 "name": "Existed_Raid", 00:09:37.935 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:37.935 "strip_size_kb": 64, 00:09:37.935 "state": "configuring", 00:09:37.935 "raid_level": "raid0", 00:09:37.935 "superblock": true, 00:09:37.935 "num_base_bdevs": 4, 00:09:37.935 "num_base_bdevs_discovered": 2, 00:09:37.935 "num_base_bdevs_operational": 4, 00:09:37.935 "base_bdevs_list": [ 00:09:37.935 { 00:09:37.935 "name": null, 00:09:37.935 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:37.935 "is_configured": false, 00:09:37.935 "data_offset": 0, 00:09:37.935 "data_size": 63488 00:09:37.935 }, 00:09:37.935 { 00:09:37.935 "name": null, 00:09:37.935 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:37.935 "is_configured": false, 00:09:37.935 "data_offset": 0, 00:09:37.935 "data_size": 63488 00:09:37.935 }, 00:09:37.935 { 00:09:37.935 "name": "BaseBdev3", 00:09:37.935 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:37.935 "is_configured": true, 00:09:37.935 "data_offset": 2048, 00:09:37.935 "data_size": 63488 00:09:37.935 }, 00:09:37.935 { 00:09:37.935 "name": "BaseBdev4", 00:09:37.935 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:37.935 "is_configured": true, 00:09:37.935 "data_offset": 2048, 00:09:37.935 "data_size": 63488 00:09:37.935 } 00:09:37.935 ] 00:09:37.935 }' 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.935 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.504 [2024-09-30 14:08:42.969264] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.504 14:08:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.504 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.504 "name": "Existed_Raid", 00:09:38.504 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:38.504 "strip_size_kb": 64, 00:09:38.504 "state": "configuring", 00:09:38.504 "raid_level": "raid0", 00:09:38.504 "superblock": true, 00:09:38.504 "num_base_bdevs": 4, 00:09:38.504 "num_base_bdevs_discovered": 3, 00:09:38.504 "num_base_bdevs_operational": 4, 00:09:38.504 "base_bdevs_list": [ 00:09:38.504 { 00:09:38.504 "name": null, 00:09:38.504 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:38.504 "is_configured": false, 00:09:38.504 "data_offset": 0, 00:09:38.504 "data_size": 63488 00:09:38.504 }, 00:09:38.504 { 00:09:38.504 "name": "BaseBdev2", 00:09:38.504 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:38.504 "is_configured": true, 00:09:38.504 "data_offset": 2048, 00:09:38.504 "data_size": 63488 00:09:38.504 }, 00:09:38.504 { 00:09:38.504 "name": "BaseBdev3", 00:09:38.504 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:38.504 "is_configured": true, 00:09:38.504 "data_offset": 2048, 00:09:38.504 "data_size": 63488 00:09:38.504 }, 00:09:38.504 { 00:09:38.504 "name": "BaseBdev4", 00:09:38.504 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:38.504 "is_configured": true, 00:09:38.504 "data_offset": 2048, 00:09:38.504 "data_size": 63488 00:09:38.504 } 00:09:38.504 ] 00:09:38.504 }' 00:09:38.504 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.504 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8e45d35a-6f33-40a1-93ba-342d91b2f552 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 [2024-09-30 14:08:43.526994] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:39.073 NewBaseBdev 00:09:39.073 [2024-09-30 14:08:43.527217] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:39.073 [2024-09-30 14:08:43.527239] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:39.073 [2024-09-30 14:08:43.527490] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:09:39.073 [2024-09-30 14:08:43.527594] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:39.073 [2024-09-30 14:08:43.527602] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:39.073 [2024-09-30 14:08:43.527693] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.073 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.073 [ 00:09:39.073 { 00:09:39.073 "name": "NewBaseBdev", 00:09:39.073 "aliases": [ 00:09:39.073 "8e45d35a-6f33-40a1-93ba-342d91b2f552" 00:09:39.073 ], 00:09:39.073 "product_name": "Malloc disk", 00:09:39.073 "block_size": 512, 00:09:39.073 "num_blocks": 65536, 00:09:39.073 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:39.073 "assigned_rate_limits": { 00:09:39.073 "rw_ios_per_sec": 0, 00:09:39.073 "rw_mbytes_per_sec": 0, 00:09:39.073 "r_mbytes_per_sec": 0, 00:09:39.073 "w_mbytes_per_sec": 0 00:09:39.073 }, 00:09:39.073 "claimed": true, 00:09:39.073 "claim_type": "exclusive_write", 00:09:39.073 "zoned": false, 00:09:39.073 "supported_io_types": { 00:09:39.073 "read": true, 00:09:39.073 "write": true, 00:09:39.074 "unmap": true, 00:09:39.074 "flush": true, 00:09:39.074 "reset": true, 00:09:39.074 "nvme_admin": false, 00:09:39.074 "nvme_io": false, 00:09:39.074 "nvme_io_md": false, 00:09:39.074 "write_zeroes": true, 00:09:39.074 "zcopy": true, 00:09:39.074 "get_zone_info": false, 00:09:39.074 "zone_management": false, 00:09:39.074 "zone_append": false, 00:09:39.074 "compare": false, 00:09:39.074 "compare_and_write": false, 00:09:39.074 "abort": true, 00:09:39.074 "seek_hole": false, 00:09:39.074 "seek_data": false, 00:09:39.074 "copy": true, 00:09:39.074 "nvme_iov_md": false 00:09:39.074 }, 00:09:39.074 "memory_domains": [ 00:09:39.074 { 00:09:39.074 "dma_device_id": "system", 00:09:39.074 "dma_device_type": 1 00:09:39.074 }, 00:09:39.074 { 00:09:39.074 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.074 "dma_device_type": 2 00:09:39.074 } 00:09:39.074 ], 00:09:39.074 "driver_specific": {} 00:09:39.074 } 00:09:39.074 ] 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.074 "name": "Existed_Raid", 00:09:39.074 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:39.074 "strip_size_kb": 64, 00:09:39.074 "state": "online", 00:09:39.074 "raid_level": "raid0", 00:09:39.074 "superblock": true, 00:09:39.074 "num_base_bdevs": 4, 00:09:39.074 "num_base_bdevs_discovered": 4, 00:09:39.074 "num_base_bdevs_operational": 4, 00:09:39.074 "base_bdevs_list": [ 00:09:39.074 { 00:09:39.074 "name": "NewBaseBdev", 00:09:39.074 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:39.074 "is_configured": true, 00:09:39.074 "data_offset": 2048, 00:09:39.074 "data_size": 63488 00:09:39.074 }, 00:09:39.074 { 00:09:39.074 "name": "BaseBdev2", 00:09:39.074 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:39.074 "is_configured": true, 00:09:39.074 "data_offset": 2048, 00:09:39.074 "data_size": 63488 00:09:39.074 }, 00:09:39.074 { 00:09:39.074 "name": "BaseBdev3", 00:09:39.074 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:39.074 "is_configured": true, 00:09:39.074 "data_offset": 2048, 00:09:39.074 "data_size": 63488 00:09:39.074 }, 00:09:39.074 { 00:09:39.074 "name": "BaseBdev4", 00:09:39.074 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:39.074 "is_configured": true, 00:09:39.074 "data_offset": 2048, 00:09:39.074 "data_size": 63488 00:09:39.074 } 00:09:39.074 ] 00:09:39.074 }' 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.074 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.334 14:08:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:39.334 [2024-09-30 14:08:43.982614] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:39.594 14:08:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.594 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:39.594 "name": "Existed_Raid", 00:09:39.594 "aliases": [ 00:09:39.594 "b2156234-6695-4d07-ab47-f008dd61b87c" 00:09:39.594 ], 00:09:39.594 "product_name": "Raid Volume", 00:09:39.594 "block_size": 512, 00:09:39.594 "num_blocks": 253952, 00:09:39.594 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:39.594 "assigned_rate_limits": { 00:09:39.594 "rw_ios_per_sec": 0, 00:09:39.594 "rw_mbytes_per_sec": 0, 00:09:39.594 "r_mbytes_per_sec": 0, 00:09:39.594 "w_mbytes_per_sec": 0 00:09:39.594 }, 00:09:39.594 "claimed": false, 00:09:39.594 "zoned": false, 00:09:39.594 "supported_io_types": { 00:09:39.594 "read": true, 00:09:39.594 "write": true, 00:09:39.594 "unmap": true, 00:09:39.594 "flush": true, 00:09:39.594 "reset": true, 00:09:39.594 "nvme_admin": false, 00:09:39.594 "nvme_io": false, 00:09:39.594 "nvme_io_md": false, 00:09:39.594 "write_zeroes": true, 00:09:39.594 "zcopy": false, 00:09:39.594 "get_zone_info": false, 00:09:39.594 "zone_management": false, 00:09:39.594 "zone_append": false, 00:09:39.594 "compare": false, 00:09:39.594 "compare_and_write": false, 00:09:39.594 "abort": false, 00:09:39.594 "seek_hole": false, 00:09:39.594 "seek_data": false, 00:09:39.594 "copy": false, 00:09:39.594 "nvme_iov_md": false 00:09:39.594 }, 00:09:39.594 "memory_domains": [ 00:09:39.594 { 00:09:39.594 "dma_device_id": "system", 00:09:39.594 "dma_device_type": 1 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.594 "dma_device_type": 2 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "system", 00:09:39.594 "dma_device_type": 1 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.594 "dma_device_type": 2 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "system", 00:09:39.594 "dma_device_type": 1 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.594 "dma_device_type": 2 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "system", 00:09:39.594 "dma_device_type": 1 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.594 "dma_device_type": 2 00:09:39.594 } 00:09:39.594 ], 00:09:39.594 "driver_specific": { 00:09:39.594 "raid": { 00:09:39.594 "uuid": "b2156234-6695-4d07-ab47-f008dd61b87c", 00:09:39.594 "strip_size_kb": 64, 00:09:39.594 "state": "online", 00:09:39.594 "raid_level": "raid0", 00:09:39.594 "superblock": true, 00:09:39.594 "num_base_bdevs": 4, 00:09:39.594 "num_base_bdevs_discovered": 4, 00:09:39.594 "num_base_bdevs_operational": 4, 00:09:39.594 "base_bdevs_list": [ 00:09:39.594 { 00:09:39.594 "name": "NewBaseBdev", 00:09:39.594 "uuid": "8e45d35a-6f33-40a1-93ba-342d91b2f552", 00:09:39.594 "is_configured": true, 00:09:39.594 "data_offset": 2048, 00:09:39.594 "data_size": 63488 00:09:39.594 }, 00:09:39.594 { 00:09:39.594 "name": "BaseBdev2", 00:09:39.594 "uuid": "abf4864d-b0b9-41e5-8873-4c024493d36f", 00:09:39.594 "is_configured": true, 00:09:39.594 "data_offset": 2048, 00:09:39.594 "data_size": 63488 00:09:39.594 }, 00:09:39.594 { 00:09:39.595 "name": "BaseBdev3", 00:09:39.595 "uuid": "2472ea7b-8bc6-483b-ba04-f17f0596ca76", 00:09:39.595 "is_configured": true, 00:09:39.595 "data_offset": 2048, 00:09:39.595 "data_size": 63488 00:09:39.595 }, 00:09:39.595 { 00:09:39.595 "name": "BaseBdev4", 00:09:39.595 "uuid": "47489534-458c-498e-9b43-a4dea536b725", 00:09:39.595 "is_configured": true, 00:09:39.595 "data_offset": 2048, 00:09:39.595 "data_size": 63488 00:09:39.595 } 00:09:39.595 ] 00:09:39.595 } 00:09:39.595 } 00:09:39.595 }' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:39.595 BaseBdev2 00:09:39.595 BaseBdev3 00:09:39.595 BaseBdev4' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.595 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.854 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.854 [2024-09-30 14:08:44.289737] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:39.854 [2024-09-30 14:08:44.289804] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:39.855 [2024-09-30 14:08:44.289906] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:39.855 [2024-09-30 14:08:44.289984] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:39.855 [2024-09-30 14:08:44.290019] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82042 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 82042 ']' 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 82042 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82042 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:39.855 killing process with pid 82042 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82042' 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 82042 00:09:39.855 [2024-09-30 14:08:44.335395] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:39.855 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 82042 00:09:39.855 [2024-09-30 14:08:44.375291] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:40.114 14:08:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:40.114 00:09:40.114 real 0m9.372s 00:09:40.114 user 0m16.035s 00:09:40.114 sys 0m1.960s 00:09:40.114 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:40.114 ************************************ 00:09:40.114 END TEST raid_state_function_test_sb 00:09:40.114 ************************************ 00:09:40.114 14:08:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.114 14:08:44 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:40.114 14:08:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:40.114 14:08:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:40.114 14:08:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:40.114 ************************************ 00:09:40.114 START TEST raid_superblock_test 00:09:40.114 ************************************ 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 4 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=82686 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 82686 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 82686 ']' 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:40.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:40.114 14:08:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.373 [2024-09-30 14:08:44.793643] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:40.373 [2024-09-30 14:08:44.793785] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82686 ] 00:09:40.373 [2024-09-30 14:08:44.931242] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:40.373 [2024-09-30 14:08:44.957062] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:40.373 [2024-09-30 14:08:45.002861] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:40.632 [2024-09-30 14:08:45.043915] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:40.632 [2024-09-30 14:08:45.044041] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 malloc1 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 [2024-09-30 14:08:45.653108] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:41.202 [2024-09-30 14:08:45.653221] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.202 [2024-09-30 14:08:45.653262] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:41.202 [2024-09-30 14:08:45.653292] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.202 [2024-09-30 14:08:45.655432] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.202 [2024-09-30 14:08:45.655512] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:41.202 pt1 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 malloc2 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 [2024-09-30 14:08:45.700245] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:41.202 [2024-09-30 14:08:45.700439] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.202 [2024-09-30 14:08:45.700548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:41.202 [2024-09-30 14:08:45.700649] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.202 [2024-09-30 14:08:45.705576] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.202 [2024-09-30 14:08:45.705643] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:41.202 pt2 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 malloc3 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 [2024-09-30 14:08:45.731216] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:41.202 [2024-09-30 14:08:45.731301] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.202 [2024-09-30 14:08:45.731338] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:41.202 [2024-09-30 14:08:45.731365] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.202 [2024-09-30 14:08:45.733358] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.202 [2024-09-30 14:08:45.733419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:41.202 pt3 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 malloc4 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 [2024-09-30 14:08:45.763434] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:41.202 [2024-09-30 14:08:45.763527] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.202 [2024-09-30 14:08:45.763563] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:41.202 [2024-09-30 14:08:45.763590] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.202 [2024-09-30 14:08:45.765560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.202 [2024-09-30 14:08:45.765620] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:41.202 pt4 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.202 [2024-09-30 14:08:45.775458] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:41.202 [2024-09-30 14:08:45.777236] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:41.202 [2024-09-30 14:08:45.777303] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:41.202 [2024-09-30 14:08:45.777360] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:41.202 [2024-09-30 14:08:45.777515] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:09:41.202 [2024-09-30 14:08:45.777526] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:41.202 [2024-09-30 14:08:45.777771] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:41.202 [2024-09-30 14:08:45.777934] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:09:41.202 [2024-09-30 14:08:45.777954] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:09:41.202 [2024-09-30 14:08:45.778068] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.202 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.203 "name": "raid_bdev1", 00:09:41.203 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:41.203 "strip_size_kb": 64, 00:09:41.203 "state": "online", 00:09:41.203 "raid_level": "raid0", 00:09:41.203 "superblock": true, 00:09:41.203 "num_base_bdevs": 4, 00:09:41.203 "num_base_bdevs_discovered": 4, 00:09:41.203 "num_base_bdevs_operational": 4, 00:09:41.203 "base_bdevs_list": [ 00:09:41.203 { 00:09:41.203 "name": "pt1", 00:09:41.203 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:41.203 "is_configured": true, 00:09:41.203 "data_offset": 2048, 00:09:41.203 "data_size": 63488 00:09:41.203 }, 00:09:41.203 { 00:09:41.203 "name": "pt2", 00:09:41.203 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:41.203 "is_configured": true, 00:09:41.203 "data_offset": 2048, 00:09:41.203 "data_size": 63488 00:09:41.203 }, 00:09:41.203 { 00:09:41.203 "name": "pt3", 00:09:41.203 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:41.203 "is_configured": true, 00:09:41.203 "data_offset": 2048, 00:09:41.203 "data_size": 63488 00:09:41.203 }, 00:09:41.203 { 00:09:41.203 "name": "pt4", 00:09:41.203 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:41.203 "is_configured": true, 00:09:41.203 "data_offset": 2048, 00:09:41.203 "data_size": 63488 00:09:41.203 } 00:09:41.203 ] 00:09:41.203 }' 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.203 14:08:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:41.772 [2024-09-30 14:08:46.230984] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.772 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:41.772 "name": "raid_bdev1", 00:09:41.772 "aliases": [ 00:09:41.772 "bdf47543-e49e-45a6-9d06-00913ab342be" 00:09:41.772 ], 00:09:41.772 "product_name": "Raid Volume", 00:09:41.772 "block_size": 512, 00:09:41.772 "num_blocks": 253952, 00:09:41.772 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:41.772 "assigned_rate_limits": { 00:09:41.772 "rw_ios_per_sec": 0, 00:09:41.772 "rw_mbytes_per_sec": 0, 00:09:41.772 "r_mbytes_per_sec": 0, 00:09:41.772 "w_mbytes_per_sec": 0 00:09:41.772 }, 00:09:41.772 "claimed": false, 00:09:41.772 "zoned": false, 00:09:41.772 "supported_io_types": { 00:09:41.772 "read": true, 00:09:41.772 "write": true, 00:09:41.772 "unmap": true, 00:09:41.772 "flush": true, 00:09:41.772 "reset": true, 00:09:41.772 "nvme_admin": false, 00:09:41.772 "nvme_io": false, 00:09:41.772 "nvme_io_md": false, 00:09:41.772 "write_zeroes": true, 00:09:41.772 "zcopy": false, 00:09:41.772 "get_zone_info": false, 00:09:41.772 "zone_management": false, 00:09:41.772 "zone_append": false, 00:09:41.772 "compare": false, 00:09:41.772 "compare_and_write": false, 00:09:41.772 "abort": false, 00:09:41.772 "seek_hole": false, 00:09:41.772 "seek_data": false, 00:09:41.772 "copy": false, 00:09:41.772 "nvme_iov_md": false 00:09:41.772 }, 00:09:41.772 "memory_domains": [ 00:09:41.772 { 00:09:41.772 "dma_device_id": "system", 00:09:41.772 "dma_device_type": 1 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.772 "dma_device_type": 2 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "system", 00:09:41.772 "dma_device_type": 1 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.772 "dma_device_type": 2 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "system", 00:09:41.772 "dma_device_type": 1 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.772 "dma_device_type": 2 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "system", 00:09:41.772 "dma_device_type": 1 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.772 "dma_device_type": 2 00:09:41.772 } 00:09:41.772 ], 00:09:41.772 "driver_specific": { 00:09:41.772 "raid": { 00:09:41.772 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:41.772 "strip_size_kb": 64, 00:09:41.772 "state": "online", 00:09:41.772 "raid_level": "raid0", 00:09:41.772 "superblock": true, 00:09:41.772 "num_base_bdevs": 4, 00:09:41.772 "num_base_bdevs_discovered": 4, 00:09:41.772 "num_base_bdevs_operational": 4, 00:09:41.772 "base_bdevs_list": [ 00:09:41.772 { 00:09:41.772 "name": "pt1", 00:09:41.772 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:41.772 "is_configured": true, 00:09:41.772 "data_offset": 2048, 00:09:41.772 "data_size": 63488 00:09:41.772 }, 00:09:41.772 { 00:09:41.772 "name": "pt2", 00:09:41.772 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:41.772 "is_configured": true, 00:09:41.772 "data_offset": 2048, 00:09:41.773 "data_size": 63488 00:09:41.773 }, 00:09:41.773 { 00:09:41.773 "name": "pt3", 00:09:41.773 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:41.773 "is_configured": true, 00:09:41.773 "data_offset": 2048, 00:09:41.773 "data_size": 63488 00:09:41.773 }, 00:09:41.773 { 00:09:41.773 "name": "pt4", 00:09:41.773 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:41.773 "is_configured": true, 00:09:41.773 "data_offset": 2048, 00:09:41.773 "data_size": 63488 00:09:41.773 } 00:09:41.773 ] 00:09:41.773 } 00:09:41.773 } 00:09:41.773 }' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:41.773 pt2 00:09:41.773 pt3 00:09:41.773 pt4' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.773 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 [2024-09-30 14:08:46.534371] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bdf47543-e49e-45a6-9d06-00913ab342be 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z bdf47543-e49e-45a6-9d06-00913ab342be ']' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 [2024-09-30 14:08:46.578032] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:42.033 [2024-09-30 14:08:46.578053] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:42.033 [2024-09-30 14:08:46.578116] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.033 [2024-09-30 14:08:46.578188] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:42.033 [2024-09-30 14:08:46.578208] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.033 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.293 [2024-09-30 14:08:46.745757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:42.293 [2024-09-30 14:08:46.747617] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:42.293 [2024-09-30 14:08:46.747694] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:42.293 [2024-09-30 14:08:46.747741] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:42.293 [2024-09-30 14:08:46.747785] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:42.293 [2024-09-30 14:08:46.747830] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:42.293 [2024-09-30 14:08:46.747848] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:42.293 [2024-09-30 14:08:46.747864] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:42.293 [2024-09-30 14:08:46.747876] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:42.293 [2024-09-30 14:08:46.747886] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:09:42.293 request: 00:09:42.293 { 00:09:42.293 "name": "raid_bdev1", 00:09:42.293 "raid_level": "raid0", 00:09:42.293 "base_bdevs": [ 00:09:42.293 "malloc1", 00:09:42.293 "malloc2", 00:09:42.293 "malloc3", 00:09:42.293 "malloc4" 00:09:42.293 ], 00:09:42.293 "strip_size_kb": 64, 00:09:42.293 "superblock": false, 00:09:42.293 "method": "bdev_raid_create", 00:09:42.293 "req_id": 1 00:09:42.293 } 00:09:42.293 Got JSON-RPC error response 00:09:42.293 response: 00:09:42.293 { 00:09:42.293 "code": -17, 00:09:42.293 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:42.293 } 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.293 [2024-09-30 14:08:46.797646] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:42.293 [2024-09-30 14:08:46.797728] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:42.293 [2024-09-30 14:08:46.797759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:42.293 [2024-09-30 14:08:46.797787] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:42.293 [2024-09-30 14:08:46.799850] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:42.293 [2024-09-30 14:08:46.799916] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:42.293 [2024-09-30 14:08:46.799999] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:42.293 [2024-09-30 14:08:46.800062] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:42.293 pt1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.293 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.293 "name": "raid_bdev1", 00:09:42.293 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:42.293 "strip_size_kb": 64, 00:09:42.294 "state": "configuring", 00:09:42.294 "raid_level": "raid0", 00:09:42.294 "superblock": true, 00:09:42.294 "num_base_bdevs": 4, 00:09:42.294 "num_base_bdevs_discovered": 1, 00:09:42.294 "num_base_bdevs_operational": 4, 00:09:42.294 "base_bdevs_list": [ 00:09:42.294 { 00:09:42.294 "name": "pt1", 00:09:42.294 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:42.294 "is_configured": true, 00:09:42.294 "data_offset": 2048, 00:09:42.294 "data_size": 63488 00:09:42.294 }, 00:09:42.294 { 00:09:42.294 "name": null, 00:09:42.294 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:42.294 "is_configured": false, 00:09:42.294 "data_offset": 2048, 00:09:42.294 "data_size": 63488 00:09:42.294 }, 00:09:42.294 { 00:09:42.294 "name": null, 00:09:42.294 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:42.294 "is_configured": false, 00:09:42.294 "data_offset": 2048, 00:09:42.294 "data_size": 63488 00:09:42.294 }, 00:09:42.294 { 00:09:42.294 "name": null, 00:09:42.294 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:42.294 "is_configured": false, 00:09:42.294 "data_offset": 2048, 00:09:42.294 "data_size": 63488 00:09:42.294 } 00:09:42.294 ] 00:09:42.294 }' 00:09:42.294 14:08:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.294 14:08:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.917 [2024-09-30 14:08:47.272900] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:42.917 [2024-09-30 14:08:47.272979] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:42.917 [2024-09-30 14:08:47.272998] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:09:42.917 [2024-09-30 14:08:47.273008] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:42.917 [2024-09-30 14:08:47.273392] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:42.917 [2024-09-30 14:08:47.273411] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:42.917 [2024-09-30 14:08:47.273482] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:42.917 [2024-09-30 14:08:47.273526] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:42.917 pt2 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.917 [2024-09-30 14:08:47.284894] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.917 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.917 "name": "raid_bdev1", 00:09:42.917 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:42.917 "strip_size_kb": 64, 00:09:42.917 "state": "configuring", 00:09:42.917 "raid_level": "raid0", 00:09:42.917 "superblock": true, 00:09:42.917 "num_base_bdevs": 4, 00:09:42.917 "num_base_bdevs_discovered": 1, 00:09:42.917 "num_base_bdevs_operational": 4, 00:09:42.917 "base_bdevs_list": [ 00:09:42.917 { 00:09:42.917 "name": "pt1", 00:09:42.917 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:42.917 "is_configured": true, 00:09:42.917 "data_offset": 2048, 00:09:42.917 "data_size": 63488 00:09:42.917 }, 00:09:42.917 { 00:09:42.917 "name": null, 00:09:42.917 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:42.917 "is_configured": false, 00:09:42.917 "data_offset": 0, 00:09:42.917 "data_size": 63488 00:09:42.917 }, 00:09:42.917 { 00:09:42.918 "name": null, 00:09:42.918 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:42.918 "is_configured": false, 00:09:42.918 "data_offset": 2048, 00:09:42.918 "data_size": 63488 00:09:42.918 }, 00:09:42.918 { 00:09:42.918 "name": null, 00:09:42.918 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:42.918 "is_configured": false, 00:09:42.918 "data_offset": 2048, 00:09:42.918 "data_size": 63488 00:09:42.918 } 00:09:42.918 ] 00:09:42.918 }' 00:09:42.918 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.918 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.177 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:43.177 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:43.177 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:43.177 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.177 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.177 [2024-09-30 14:08:47.720163] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:43.177 [2024-09-30 14:08:47.720296] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:43.177 [2024-09-30 14:08:47.720334] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:09:43.177 [2024-09-30 14:08:47.720361] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:43.178 [2024-09-30 14:08:47.720784] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:43.178 [2024-09-30 14:08:47.720846] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:43.178 [2024-09-30 14:08:47.720948] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:43.178 [2024-09-30 14:08:47.721010] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:43.178 pt2 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.178 [2024-09-30 14:08:47.732099] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:43.178 [2024-09-30 14:08:47.732184] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:43.178 [2024-09-30 14:08:47.732219] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:09:43.178 [2024-09-30 14:08:47.732245] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:43.178 [2024-09-30 14:08:47.732615] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:43.178 [2024-09-30 14:08:47.732667] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:43.178 [2024-09-30 14:08:47.732753] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:43.178 [2024-09-30 14:08:47.732798] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:43.178 pt3 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.178 [2024-09-30 14:08:47.744068] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:43.178 [2024-09-30 14:08:47.744153] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:43.178 [2024-09-30 14:08:47.744189] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:09:43.178 [2024-09-30 14:08:47.744198] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:43.178 [2024-09-30 14:08:47.744535] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:43.178 [2024-09-30 14:08:47.744552] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:43.178 [2024-09-30 14:08:47.744609] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:43.178 [2024-09-30 14:08:47.744632] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:43.178 [2024-09-30 14:08:47.744736] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:43.178 [2024-09-30 14:08:47.744743] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:43.178 [2024-09-30 14:08:47.744963] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:43.178 [2024-09-30 14:08:47.745070] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:43.178 [2024-09-30 14:08:47.745083] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:43.178 [2024-09-30 14:08:47.745174] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:43.178 pt4 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.178 "name": "raid_bdev1", 00:09:43.178 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:43.178 "strip_size_kb": 64, 00:09:43.178 "state": "online", 00:09:43.178 "raid_level": "raid0", 00:09:43.178 "superblock": true, 00:09:43.178 "num_base_bdevs": 4, 00:09:43.178 "num_base_bdevs_discovered": 4, 00:09:43.178 "num_base_bdevs_operational": 4, 00:09:43.178 "base_bdevs_list": [ 00:09:43.178 { 00:09:43.178 "name": "pt1", 00:09:43.178 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:43.178 "is_configured": true, 00:09:43.178 "data_offset": 2048, 00:09:43.178 "data_size": 63488 00:09:43.178 }, 00:09:43.178 { 00:09:43.178 "name": "pt2", 00:09:43.178 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:43.178 "is_configured": true, 00:09:43.178 "data_offset": 2048, 00:09:43.178 "data_size": 63488 00:09:43.178 }, 00:09:43.178 { 00:09:43.178 "name": "pt3", 00:09:43.178 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:43.178 "is_configured": true, 00:09:43.178 "data_offset": 2048, 00:09:43.178 "data_size": 63488 00:09:43.178 }, 00:09:43.178 { 00:09:43.178 "name": "pt4", 00:09:43.178 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:43.178 "is_configured": true, 00:09:43.178 "data_offset": 2048, 00:09:43.178 "data_size": 63488 00:09:43.178 } 00:09:43.178 ] 00:09:43.178 }' 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.178 14:08:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.748 [2024-09-30 14:08:48.171694] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:43.748 "name": "raid_bdev1", 00:09:43.748 "aliases": [ 00:09:43.748 "bdf47543-e49e-45a6-9d06-00913ab342be" 00:09:43.748 ], 00:09:43.748 "product_name": "Raid Volume", 00:09:43.748 "block_size": 512, 00:09:43.748 "num_blocks": 253952, 00:09:43.748 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:43.748 "assigned_rate_limits": { 00:09:43.748 "rw_ios_per_sec": 0, 00:09:43.748 "rw_mbytes_per_sec": 0, 00:09:43.748 "r_mbytes_per_sec": 0, 00:09:43.748 "w_mbytes_per_sec": 0 00:09:43.748 }, 00:09:43.748 "claimed": false, 00:09:43.748 "zoned": false, 00:09:43.748 "supported_io_types": { 00:09:43.748 "read": true, 00:09:43.748 "write": true, 00:09:43.748 "unmap": true, 00:09:43.748 "flush": true, 00:09:43.748 "reset": true, 00:09:43.748 "nvme_admin": false, 00:09:43.748 "nvme_io": false, 00:09:43.748 "nvme_io_md": false, 00:09:43.748 "write_zeroes": true, 00:09:43.748 "zcopy": false, 00:09:43.748 "get_zone_info": false, 00:09:43.748 "zone_management": false, 00:09:43.748 "zone_append": false, 00:09:43.748 "compare": false, 00:09:43.748 "compare_and_write": false, 00:09:43.748 "abort": false, 00:09:43.748 "seek_hole": false, 00:09:43.748 "seek_data": false, 00:09:43.748 "copy": false, 00:09:43.748 "nvme_iov_md": false 00:09:43.748 }, 00:09:43.748 "memory_domains": [ 00:09:43.748 { 00:09:43.748 "dma_device_id": "system", 00:09:43.748 "dma_device_type": 1 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.748 "dma_device_type": 2 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "system", 00:09:43.748 "dma_device_type": 1 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.748 "dma_device_type": 2 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "system", 00:09:43.748 "dma_device_type": 1 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.748 "dma_device_type": 2 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "system", 00:09:43.748 "dma_device_type": 1 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.748 "dma_device_type": 2 00:09:43.748 } 00:09:43.748 ], 00:09:43.748 "driver_specific": { 00:09:43.748 "raid": { 00:09:43.748 "uuid": "bdf47543-e49e-45a6-9d06-00913ab342be", 00:09:43.748 "strip_size_kb": 64, 00:09:43.748 "state": "online", 00:09:43.748 "raid_level": "raid0", 00:09:43.748 "superblock": true, 00:09:43.748 "num_base_bdevs": 4, 00:09:43.748 "num_base_bdevs_discovered": 4, 00:09:43.748 "num_base_bdevs_operational": 4, 00:09:43.748 "base_bdevs_list": [ 00:09:43.748 { 00:09:43.748 "name": "pt1", 00:09:43.748 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:43.748 "is_configured": true, 00:09:43.748 "data_offset": 2048, 00:09:43.748 "data_size": 63488 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "name": "pt2", 00:09:43.748 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:43.748 "is_configured": true, 00:09:43.748 "data_offset": 2048, 00:09:43.748 "data_size": 63488 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "name": "pt3", 00:09:43.748 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:43.748 "is_configured": true, 00:09:43.748 "data_offset": 2048, 00:09:43.748 "data_size": 63488 00:09:43.748 }, 00:09:43.748 { 00:09:43.748 "name": "pt4", 00:09:43.748 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:43.748 "is_configured": true, 00:09:43.748 "data_offset": 2048, 00:09:43.748 "data_size": 63488 00:09:43.748 } 00:09:43.748 ] 00:09:43.748 } 00:09:43.748 } 00:09:43.748 }' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:43.748 pt2 00:09:43.748 pt3 00:09:43.748 pt4' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.748 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.008 [2024-09-30 14:08:48.475053] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.008 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' bdf47543-e49e-45a6-9d06-00913ab342be '!=' bdf47543-e49e-45a6-9d06-00913ab342be ']' 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 82686 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 82686 ']' 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 82686 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82686 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82686' 00:09:44.009 killing process with pid 82686 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 82686 00:09:44.009 [2024-09-30 14:08:48.561800] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:44.009 [2024-09-30 14:08:48.561925] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:44.009 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 82686 00:09:44.009 [2024-09-30 14:08:48.562014] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:44.009 [2024-09-30 14:08:48.562023] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:44.009 [2024-09-30 14:08:48.605314] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:44.268 14:08:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:44.268 00:09:44.268 real 0m4.156s 00:09:44.268 user 0m6.517s 00:09:44.268 sys 0m0.954s 00:09:44.268 ************************************ 00:09:44.268 END TEST raid_superblock_test 00:09:44.268 ************************************ 00:09:44.268 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:44.268 14:08:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.268 14:08:48 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:09:44.268 14:08:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:44.268 14:08:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:44.268 14:08:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:44.531 ************************************ 00:09:44.531 START TEST raid_read_error_test 00:09:44.531 ************************************ 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 read 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.AcsZQp7hYS 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82934 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82934 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 82934 ']' 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:44.531 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:44.531 14:08:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.531 [2024-09-30 14:08:49.041543] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:44.531 [2024-09-30 14:08:49.041679] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82934 ] 00:09:44.531 [2024-09-30 14:08:49.178538] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:44.796 [2024-09-30 14:08:49.209269] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:44.796 [2024-09-30 14:08:49.255262] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:44.796 [2024-09-30 14:08:49.296558] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.796 [2024-09-30 14:08:49.296593] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 BaseBdev1_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 true 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 [2024-09-30 14:08:49.922119] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:45.364 [2024-09-30 14:08:49.922213] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.364 [2024-09-30 14:08:49.922251] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:45.364 [2024-09-30 14:08:49.922264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.364 [2024-09-30 14:08:49.924344] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.364 [2024-09-30 14:08:49.924383] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:45.364 BaseBdev1 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 BaseBdev2_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 true 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 [2024-09-30 14:08:49.969708] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:45.364 [2024-09-30 14:08:49.969794] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.364 [2024-09-30 14:08:49.969828] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:45.364 [2024-09-30 14:08:49.969838] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.364 [2024-09-30 14:08:49.971872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.364 [2024-09-30 14:08:49.971911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:45.364 BaseBdev2 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 BaseBdev3_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.364 14:08:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.364 true 00:09:45.364 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.364 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:45.364 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.365 [2024-09-30 14:08:50.009864] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:45.365 [2024-09-30 14:08:50.009912] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.365 [2024-09-30 14:08:50.009927] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:45.365 [2024-09-30 14:08:50.009936] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.365 [2024-09-30 14:08:50.011990] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.365 [2024-09-30 14:08:50.012075] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:45.365 BaseBdev3 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.365 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.624 BaseBdev4_malloc 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.624 true 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.624 [2024-09-30 14:08:50.050149] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:45.624 [2024-09-30 14:08:50.050238] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.624 [2024-09-30 14:08:50.050257] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:45.624 [2024-09-30 14:08:50.050267] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.624 [2024-09-30 14:08:50.052263] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.624 [2024-09-30 14:08:50.052303] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:45.624 BaseBdev4 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.624 [2024-09-30 14:08:50.062196] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:45.624 [2024-09-30 14:08:50.063986] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:45.624 [2024-09-30 14:08:50.064055] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:45.624 [2024-09-30 14:08:50.064118] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:45.624 [2024-09-30 14:08:50.064292] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:09:45.624 [2024-09-30 14:08:50.064305] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:45.624 [2024-09-30 14:08:50.064560] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:45.624 [2024-09-30 14:08:50.064683] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:09:45.624 [2024-09-30 14:08:50.064693] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:09:45.624 [2024-09-30 14:08:50.064805] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.624 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.624 "name": "raid_bdev1", 00:09:45.624 "uuid": "e986d579-9bc4-433a-8024-8465cae7ff6b", 00:09:45.624 "strip_size_kb": 64, 00:09:45.624 "state": "online", 00:09:45.624 "raid_level": "raid0", 00:09:45.624 "superblock": true, 00:09:45.624 "num_base_bdevs": 4, 00:09:45.624 "num_base_bdevs_discovered": 4, 00:09:45.624 "num_base_bdevs_operational": 4, 00:09:45.624 "base_bdevs_list": [ 00:09:45.624 { 00:09:45.624 "name": "BaseBdev1", 00:09:45.624 "uuid": "e40d08bf-dd6a-5e80-b206-0a489afd2993", 00:09:45.624 "is_configured": true, 00:09:45.624 "data_offset": 2048, 00:09:45.624 "data_size": 63488 00:09:45.624 }, 00:09:45.624 { 00:09:45.624 "name": "BaseBdev2", 00:09:45.624 "uuid": "47a08de5-a698-5135-a8ed-b5a464226bc9", 00:09:45.624 "is_configured": true, 00:09:45.624 "data_offset": 2048, 00:09:45.624 "data_size": 63488 00:09:45.624 }, 00:09:45.624 { 00:09:45.624 "name": "BaseBdev3", 00:09:45.624 "uuid": "98004123-1bc6-5026-8ac4-db915799c930", 00:09:45.624 "is_configured": true, 00:09:45.624 "data_offset": 2048, 00:09:45.625 "data_size": 63488 00:09:45.625 }, 00:09:45.625 { 00:09:45.625 "name": "BaseBdev4", 00:09:45.625 "uuid": "2385b5ae-205b-5af2-9424-c990c5cc3e28", 00:09:45.625 "is_configured": true, 00:09:45.625 "data_offset": 2048, 00:09:45.625 "data_size": 63488 00:09:45.625 } 00:09:45.625 ] 00:09:45.625 }' 00:09:45.625 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.625 14:08:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.884 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:45.884 14:08:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:46.143 [2024-09-30 14:08:50.557693] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:09:47.080 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:47.080 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.080 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.080 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.080 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.081 "name": "raid_bdev1", 00:09:47.081 "uuid": "e986d579-9bc4-433a-8024-8465cae7ff6b", 00:09:47.081 "strip_size_kb": 64, 00:09:47.081 "state": "online", 00:09:47.081 "raid_level": "raid0", 00:09:47.081 "superblock": true, 00:09:47.081 "num_base_bdevs": 4, 00:09:47.081 "num_base_bdevs_discovered": 4, 00:09:47.081 "num_base_bdevs_operational": 4, 00:09:47.081 "base_bdevs_list": [ 00:09:47.081 { 00:09:47.081 "name": "BaseBdev1", 00:09:47.081 "uuid": "e40d08bf-dd6a-5e80-b206-0a489afd2993", 00:09:47.081 "is_configured": true, 00:09:47.081 "data_offset": 2048, 00:09:47.081 "data_size": 63488 00:09:47.081 }, 00:09:47.081 { 00:09:47.081 "name": "BaseBdev2", 00:09:47.081 "uuid": "47a08de5-a698-5135-a8ed-b5a464226bc9", 00:09:47.081 "is_configured": true, 00:09:47.081 "data_offset": 2048, 00:09:47.081 "data_size": 63488 00:09:47.081 }, 00:09:47.081 { 00:09:47.081 "name": "BaseBdev3", 00:09:47.081 "uuid": "98004123-1bc6-5026-8ac4-db915799c930", 00:09:47.081 "is_configured": true, 00:09:47.081 "data_offset": 2048, 00:09:47.081 "data_size": 63488 00:09:47.081 }, 00:09:47.081 { 00:09:47.081 "name": "BaseBdev4", 00:09:47.081 "uuid": "2385b5ae-205b-5af2-9424-c990c5cc3e28", 00:09:47.081 "is_configured": true, 00:09:47.081 "data_offset": 2048, 00:09:47.081 "data_size": 63488 00:09:47.081 } 00:09:47.081 ] 00:09:47.081 }' 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.081 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.340 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:47.340 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.340 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.340 [2024-09-30 14:08:51.933090] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:47.340 [2024-09-30 14:08:51.933126] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:47.340 [2024-09-30 14:08:51.935646] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:47.340 [2024-09-30 14:08:51.935730] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.340 [2024-09-30 14:08:51.935802] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:47.340 [2024-09-30 14:08:51.935848] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:09:47.340 { 00:09:47.340 "results": [ 00:09:47.340 { 00:09:47.340 "job": "raid_bdev1", 00:09:47.340 "core_mask": "0x1", 00:09:47.341 "workload": "randrw", 00:09:47.341 "percentage": 50, 00:09:47.341 "status": "finished", 00:09:47.341 "queue_depth": 1, 00:09:47.341 "io_size": 131072, 00:09:47.341 "runtime": 1.37626, 00:09:47.341 "iops": 17142.109775769113, 00:09:47.341 "mibps": 2142.763721971139, 00:09:47.341 "io_failed": 1, 00:09:47.341 "io_timeout": 0, 00:09:47.341 "avg_latency_us": 80.95807382731574, 00:09:47.341 "min_latency_us": 24.705676855895195, 00:09:47.341 "max_latency_us": 1323.598253275109 00:09:47.341 } 00:09:47.341 ], 00:09:47.341 "core_count": 1 00:09:47.341 } 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82934 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 82934 ']' 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 82934 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82934 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82934' 00:09:47.341 killing process with pid 82934 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 82934 00:09:47.341 [2024-09-30 14:08:51.981436] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:47.341 14:08:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 82934 00:09:47.601 [2024-09-30 14:08:52.016438] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:47.601 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.AcsZQp7hYS 00:09:47.601 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:47.601 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:47.860 00:09:47.860 real 0m3.334s 00:09:47.860 user 0m4.178s 00:09:47.860 sys 0m0.565s 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:47.860 14:08:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.860 ************************************ 00:09:47.860 END TEST raid_read_error_test 00:09:47.860 ************************************ 00:09:47.860 14:08:52 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:09:47.860 14:08:52 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:47.860 14:08:52 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:47.860 14:08:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:47.861 ************************************ 00:09:47.861 START TEST raid_write_error_test 00:09:47.861 ************************************ 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 write 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Nr8OOeM97u 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83069 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83069 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 83069 ']' 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:47.861 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:47.861 14:08:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.861 [2024-09-30 14:08:52.445695] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:47.861 [2024-09-30 14:08:52.445866] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83069 ] 00:09:48.120 [2024-09-30 14:08:52.577211] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:48.120 [2024-09-30 14:08:52.605767] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:48.120 [2024-09-30 14:08:52.650824] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:48.120 [2024-09-30 14:08:52.691763] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:48.120 [2024-09-30 14:08:52.691799] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.689 BaseBdev1_malloc 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.689 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.690 true 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.690 [2024-09-30 14:08:53.317092] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:48.690 [2024-09-30 14:08:53.317155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.690 [2024-09-30 14:08:53.317173] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:48.690 [2024-09-30 14:08:53.317186] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.690 [2024-09-30 14:08:53.319284] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.690 [2024-09-30 14:08:53.319389] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:48.690 BaseBdev1 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.690 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.949 BaseBdev2_malloc 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.949 true 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.949 [2024-09-30 14:08:53.375517] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:48.949 [2024-09-30 14:08:53.375659] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.949 [2024-09-30 14:08:53.375691] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:48.949 [2024-09-30 14:08:53.375710] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.949 [2024-09-30 14:08:53.379068] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.949 [2024-09-30 14:08:53.379189] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:48.949 BaseBdev2 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.949 BaseBdev3_malloc 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.949 true 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:48.949 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 [2024-09-30 14:08:53.416096] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:48.950 [2024-09-30 14:08:53.416153] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.950 [2024-09-30 14:08:53.416184] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:48.950 [2024-09-30 14:08:53.416193] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.950 [2024-09-30 14:08:53.418150] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.950 [2024-09-30 14:08:53.418223] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:48.950 BaseBdev3 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 BaseBdev4_malloc 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 true 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 [2024-09-30 14:08:53.456293] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:48.950 [2024-09-30 14:08:53.456342] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.950 [2024-09-30 14:08:53.456359] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:48.950 [2024-09-30 14:08:53.456369] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.950 [2024-09-30 14:08:53.458447] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.950 [2024-09-30 14:08:53.458496] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:48.950 BaseBdev4 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 [2024-09-30 14:08:53.468339] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:48.950 [2024-09-30 14:08:53.470099] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:48.950 [2024-09-30 14:08:53.470164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:48.950 [2024-09-30 14:08:53.470216] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:48.950 [2024-09-30 14:08:53.470395] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:09:48.950 [2024-09-30 14:08:53.470409] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:48.950 [2024-09-30 14:08:53.470647] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:48.950 [2024-09-30 14:08:53.470781] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:09:48.950 [2024-09-30 14:08:53.470798] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:09:48.950 [2024-09-30 14:08:53.470918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.950 "name": "raid_bdev1", 00:09:48.950 "uuid": "df859878-fd78-40ab-bdcc-736cecdeb76d", 00:09:48.950 "strip_size_kb": 64, 00:09:48.950 "state": "online", 00:09:48.950 "raid_level": "raid0", 00:09:48.950 "superblock": true, 00:09:48.950 "num_base_bdevs": 4, 00:09:48.950 "num_base_bdevs_discovered": 4, 00:09:48.950 "num_base_bdevs_operational": 4, 00:09:48.950 "base_bdevs_list": [ 00:09:48.950 { 00:09:48.950 "name": "BaseBdev1", 00:09:48.950 "uuid": "53bfbbd2-6850-5c30-bb2d-d92bf45505e3", 00:09:48.950 "is_configured": true, 00:09:48.950 "data_offset": 2048, 00:09:48.950 "data_size": 63488 00:09:48.950 }, 00:09:48.950 { 00:09:48.950 "name": "BaseBdev2", 00:09:48.950 "uuid": "cf292156-ec73-5c35-aa18-51ad3651b360", 00:09:48.950 "is_configured": true, 00:09:48.950 "data_offset": 2048, 00:09:48.950 "data_size": 63488 00:09:48.950 }, 00:09:48.950 { 00:09:48.950 "name": "BaseBdev3", 00:09:48.950 "uuid": "e4d86525-fb5a-51d4-917d-d54cfe242787", 00:09:48.950 "is_configured": true, 00:09:48.950 "data_offset": 2048, 00:09:48.950 "data_size": 63488 00:09:48.950 }, 00:09:48.950 { 00:09:48.950 "name": "BaseBdev4", 00:09:48.950 "uuid": "1df7615d-af31-55ac-a3e6-5b8cc81264c2", 00:09:48.950 "is_configured": true, 00:09:48.950 "data_offset": 2048, 00:09:48.950 "data_size": 63488 00:09:48.950 } 00:09:48.950 ] 00:09:48.950 }' 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.950 14:08:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.519 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:49.519 14:08:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:49.519 [2024-09-30 14:08:54.015798] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.459 14:08:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.459 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.459 "name": "raid_bdev1", 00:09:50.459 "uuid": "df859878-fd78-40ab-bdcc-736cecdeb76d", 00:09:50.459 "strip_size_kb": 64, 00:09:50.459 "state": "online", 00:09:50.459 "raid_level": "raid0", 00:09:50.459 "superblock": true, 00:09:50.459 "num_base_bdevs": 4, 00:09:50.459 "num_base_bdevs_discovered": 4, 00:09:50.459 "num_base_bdevs_operational": 4, 00:09:50.459 "base_bdevs_list": [ 00:09:50.459 { 00:09:50.459 "name": "BaseBdev1", 00:09:50.459 "uuid": "53bfbbd2-6850-5c30-bb2d-d92bf45505e3", 00:09:50.459 "is_configured": true, 00:09:50.459 "data_offset": 2048, 00:09:50.459 "data_size": 63488 00:09:50.459 }, 00:09:50.459 { 00:09:50.459 "name": "BaseBdev2", 00:09:50.459 "uuid": "cf292156-ec73-5c35-aa18-51ad3651b360", 00:09:50.459 "is_configured": true, 00:09:50.459 "data_offset": 2048, 00:09:50.459 "data_size": 63488 00:09:50.459 }, 00:09:50.459 { 00:09:50.459 "name": "BaseBdev3", 00:09:50.459 "uuid": "e4d86525-fb5a-51d4-917d-d54cfe242787", 00:09:50.459 "is_configured": true, 00:09:50.459 "data_offset": 2048, 00:09:50.459 "data_size": 63488 00:09:50.459 }, 00:09:50.459 { 00:09:50.459 "name": "BaseBdev4", 00:09:50.459 "uuid": "1df7615d-af31-55ac-a3e6-5b8cc81264c2", 00:09:50.459 "is_configured": true, 00:09:50.459 "data_offset": 2048, 00:09:50.459 "data_size": 63488 00:09:50.459 } 00:09:50.459 ] 00:09:50.459 }' 00:09:50.459 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.459 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.029 [2024-09-30 14:08:55.403312] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:51.029 [2024-09-30 14:08:55.403413] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:51.029 [2024-09-30 14:08:55.405945] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:51.029 [2024-09-30 14:08:55.406036] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:51.029 [2024-09-30 14:08:55.406098] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:51.029 [2024-09-30 14:08:55.406140] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:09:51.029 { 00:09:51.029 "results": [ 00:09:51.029 { 00:09:51.029 "job": "raid_bdev1", 00:09:51.029 "core_mask": "0x1", 00:09:51.029 "workload": "randrw", 00:09:51.029 "percentage": 50, 00:09:51.029 "status": "finished", 00:09:51.029 "queue_depth": 1, 00:09:51.029 "io_size": 131072, 00:09:51.029 "runtime": 1.388506, 00:09:51.029 "iops": 17024.773389528025, 00:09:51.029 "mibps": 2128.096673691003, 00:09:51.029 "io_failed": 1, 00:09:51.029 "io_timeout": 0, 00:09:51.029 "avg_latency_us": 81.5737689801166, 00:09:51.029 "min_latency_us": 24.817467248908297, 00:09:51.029 "max_latency_us": 1373.6803493449781 00:09:51.029 } 00:09:51.029 ], 00:09:51.029 "core_count": 1 00:09:51.029 } 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83069 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 83069 ']' 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 83069 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83069 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83069' 00:09:51.029 killing process with pid 83069 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 83069 00:09:51.029 [2024-09-30 14:08:55.457662] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:51.029 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 83069 00:09:51.029 [2024-09-30 14:08:55.491949] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Nr8OOeM97u 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:51.289 ************************************ 00:09:51.289 END TEST raid_write_error_test 00:09:51.289 ************************************ 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:51.289 00:09:51.289 real 0m3.397s 00:09:51.289 user 0m4.235s 00:09:51.289 sys 0m0.587s 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:51.289 14:08:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.289 14:08:55 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:51.290 14:08:55 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:09:51.290 14:08:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:51.290 14:08:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:51.290 14:08:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:51.290 ************************************ 00:09:51.290 START TEST raid_state_function_test 00:09:51.290 ************************************ 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 false 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:51.290 Process raid pid: 83201 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=83201 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83201' 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 83201 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 83201 ']' 00:09:51.290 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:51.290 14:08:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.290 [2024-09-30 14:08:55.911524] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:09:51.290 [2024-09-30 14:08:55.911657] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:51.550 [2024-09-30 14:08:56.048866] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:09:51.550 [2024-09-30 14:08:56.064051] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:51.550 [2024-09-30 14:08:56.110948] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:51.550 [2024-09-30 14:08:56.152147] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:51.550 [2024-09-30 14:08:56.152177] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:52.118 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.119 [2024-09-30 14:08:56.729196] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:52.119 [2024-09-30 14:08:56.729318] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:52.119 [2024-09-30 14:08:56.729334] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:52.119 [2024-09-30 14:08:56.729342] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:52.119 [2024-09-30 14:08:56.729352] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:52.119 [2024-09-30 14:08:56.729358] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:52.119 [2024-09-30 14:08:56.729365] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:52.119 [2024-09-30 14:08:56.729372] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.119 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.378 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.378 "name": "Existed_Raid", 00:09:52.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.378 "strip_size_kb": 64, 00:09:52.378 "state": "configuring", 00:09:52.378 "raid_level": "concat", 00:09:52.378 "superblock": false, 00:09:52.378 "num_base_bdevs": 4, 00:09:52.378 "num_base_bdevs_discovered": 0, 00:09:52.378 "num_base_bdevs_operational": 4, 00:09:52.378 "base_bdevs_list": [ 00:09:52.378 { 00:09:52.378 "name": "BaseBdev1", 00:09:52.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.378 "is_configured": false, 00:09:52.378 "data_offset": 0, 00:09:52.378 "data_size": 0 00:09:52.378 }, 00:09:52.378 { 00:09:52.378 "name": "BaseBdev2", 00:09:52.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.378 "is_configured": false, 00:09:52.378 "data_offset": 0, 00:09:52.378 "data_size": 0 00:09:52.378 }, 00:09:52.378 { 00:09:52.378 "name": "BaseBdev3", 00:09:52.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.378 "is_configured": false, 00:09:52.378 "data_offset": 0, 00:09:52.378 "data_size": 0 00:09:52.378 }, 00:09:52.378 { 00:09:52.378 "name": "BaseBdev4", 00:09:52.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.378 "is_configured": false, 00:09:52.378 "data_offset": 0, 00:09:52.378 "data_size": 0 00:09:52.378 } 00:09:52.378 ] 00:09:52.378 }' 00:09:52.378 14:08:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.378 14:08:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 [2024-09-30 14:08:57.188371] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:52.637 [2024-09-30 14:08:57.188468] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 [2024-09-30 14:08:57.200341] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:52.637 [2024-09-30 14:08:57.200411] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:52.637 [2024-09-30 14:08:57.200438] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:52.637 [2024-09-30 14:08:57.200458] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:52.637 [2024-09-30 14:08:57.200476] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:52.637 [2024-09-30 14:08:57.200501] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:52.637 [2024-09-30 14:08:57.200520] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:52.637 [2024-09-30 14:08:57.200537] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 [2024-09-30 14:08:57.220817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:52.637 BaseBdev1 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.637 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.637 [ 00:09:52.637 { 00:09:52.637 "name": "BaseBdev1", 00:09:52.637 "aliases": [ 00:09:52.637 "c4141592-4ace-4a12-8013-a97fda723f75" 00:09:52.637 ], 00:09:52.637 "product_name": "Malloc disk", 00:09:52.638 "block_size": 512, 00:09:52.638 "num_blocks": 65536, 00:09:52.638 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:52.638 "assigned_rate_limits": { 00:09:52.638 "rw_ios_per_sec": 0, 00:09:52.638 "rw_mbytes_per_sec": 0, 00:09:52.638 "r_mbytes_per_sec": 0, 00:09:52.638 "w_mbytes_per_sec": 0 00:09:52.638 }, 00:09:52.638 "claimed": true, 00:09:52.638 "claim_type": "exclusive_write", 00:09:52.638 "zoned": false, 00:09:52.638 "supported_io_types": { 00:09:52.638 "read": true, 00:09:52.638 "write": true, 00:09:52.638 "unmap": true, 00:09:52.638 "flush": true, 00:09:52.638 "reset": true, 00:09:52.638 "nvme_admin": false, 00:09:52.638 "nvme_io": false, 00:09:52.638 "nvme_io_md": false, 00:09:52.638 "write_zeroes": true, 00:09:52.638 "zcopy": true, 00:09:52.638 "get_zone_info": false, 00:09:52.638 "zone_management": false, 00:09:52.638 "zone_append": false, 00:09:52.638 "compare": false, 00:09:52.638 "compare_and_write": false, 00:09:52.638 "abort": true, 00:09:52.638 "seek_hole": false, 00:09:52.638 "seek_data": false, 00:09:52.638 "copy": true, 00:09:52.638 "nvme_iov_md": false 00:09:52.638 }, 00:09:52.638 "memory_domains": [ 00:09:52.638 { 00:09:52.638 "dma_device_id": "system", 00:09:52.638 "dma_device_type": 1 00:09:52.638 }, 00:09:52.638 { 00:09:52.638 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.638 "dma_device_type": 2 00:09:52.638 } 00:09:52.638 ], 00:09:52.638 "driver_specific": {} 00:09:52.638 } 00:09:52.638 ] 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.638 "name": "Existed_Raid", 00:09:52.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.638 "strip_size_kb": 64, 00:09:52.638 "state": "configuring", 00:09:52.638 "raid_level": "concat", 00:09:52.638 "superblock": false, 00:09:52.638 "num_base_bdevs": 4, 00:09:52.638 "num_base_bdevs_discovered": 1, 00:09:52.638 "num_base_bdevs_operational": 4, 00:09:52.638 "base_bdevs_list": [ 00:09:52.638 { 00:09:52.638 "name": "BaseBdev1", 00:09:52.638 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:52.638 "is_configured": true, 00:09:52.638 "data_offset": 0, 00:09:52.638 "data_size": 65536 00:09:52.638 }, 00:09:52.638 { 00:09:52.638 "name": "BaseBdev2", 00:09:52.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.638 "is_configured": false, 00:09:52.638 "data_offset": 0, 00:09:52.638 "data_size": 0 00:09:52.638 }, 00:09:52.638 { 00:09:52.638 "name": "BaseBdev3", 00:09:52.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.638 "is_configured": false, 00:09:52.638 "data_offset": 0, 00:09:52.638 "data_size": 0 00:09:52.638 }, 00:09:52.638 { 00:09:52.638 "name": "BaseBdev4", 00:09:52.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.638 "is_configured": false, 00:09:52.638 "data_offset": 0, 00:09:52.638 "data_size": 0 00:09:52.638 } 00:09:52.638 ] 00:09:52.638 }' 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.638 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.207 [2024-09-30 14:08:57.620164] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:53.207 [2024-09-30 14:08:57.620248] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.207 [2024-09-30 14:08:57.628179] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:53.207 [2024-09-30 14:08:57.629982] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:53.207 [2024-09-30 14:08:57.630051] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:53.207 [2024-09-30 14:08:57.630095] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:53.207 [2024-09-30 14:08:57.630117] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:53.207 [2024-09-30 14:08:57.630136] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:53.207 [2024-09-30 14:08:57.630154] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.207 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.208 "name": "Existed_Raid", 00:09:53.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.208 "strip_size_kb": 64, 00:09:53.208 "state": "configuring", 00:09:53.208 "raid_level": "concat", 00:09:53.208 "superblock": false, 00:09:53.208 "num_base_bdevs": 4, 00:09:53.208 "num_base_bdevs_discovered": 1, 00:09:53.208 "num_base_bdevs_operational": 4, 00:09:53.208 "base_bdevs_list": [ 00:09:53.208 { 00:09:53.208 "name": "BaseBdev1", 00:09:53.208 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:53.208 "is_configured": true, 00:09:53.208 "data_offset": 0, 00:09:53.208 "data_size": 65536 00:09:53.208 }, 00:09:53.208 { 00:09:53.208 "name": "BaseBdev2", 00:09:53.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.208 "is_configured": false, 00:09:53.208 "data_offset": 0, 00:09:53.208 "data_size": 0 00:09:53.208 }, 00:09:53.208 { 00:09:53.208 "name": "BaseBdev3", 00:09:53.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.208 "is_configured": false, 00:09:53.208 "data_offset": 0, 00:09:53.208 "data_size": 0 00:09:53.208 }, 00:09:53.208 { 00:09:53.208 "name": "BaseBdev4", 00:09:53.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.208 "is_configured": false, 00:09:53.208 "data_offset": 0, 00:09:53.208 "data_size": 0 00:09:53.208 } 00:09:53.208 ] 00:09:53.208 }' 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.208 14:08:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.467 [2024-09-30 14:08:58.082039] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:53.467 BaseBdev2 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.467 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.467 [ 00:09:53.467 { 00:09:53.467 "name": "BaseBdev2", 00:09:53.467 "aliases": [ 00:09:53.467 "eaf87b3e-7905-4426-8782-151dccf84103" 00:09:53.467 ], 00:09:53.467 "product_name": "Malloc disk", 00:09:53.467 "block_size": 512, 00:09:53.467 "num_blocks": 65536, 00:09:53.467 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:53.467 "assigned_rate_limits": { 00:09:53.467 "rw_ios_per_sec": 0, 00:09:53.467 "rw_mbytes_per_sec": 0, 00:09:53.467 "r_mbytes_per_sec": 0, 00:09:53.467 "w_mbytes_per_sec": 0 00:09:53.467 }, 00:09:53.467 "claimed": true, 00:09:53.467 "claim_type": "exclusive_write", 00:09:53.467 "zoned": false, 00:09:53.467 "supported_io_types": { 00:09:53.467 "read": true, 00:09:53.467 "write": true, 00:09:53.467 "unmap": true, 00:09:53.467 "flush": true, 00:09:53.467 "reset": true, 00:09:53.467 "nvme_admin": false, 00:09:53.467 "nvme_io": false, 00:09:53.467 "nvme_io_md": false, 00:09:53.467 "write_zeroes": true, 00:09:53.467 "zcopy": true, 00:09:53.467 "get_zone_info": false, 00:09:53.467 "zone_management": false, 00:09:53.467 "zone_append": false, 00:09:53.467 "compare": false, 00:09:53.467 "compare_and_write": false, 00:09:53.467 "abort": true, 00:09:53.467 "seek_hole": false, 00:09:53.467 "seek_data": false, 00:09:53.467 "copy": true, 00:09:53.467 "nvme_iov_md": false 00:09:53.467 }, 00:09:53.467 "memory_domains": [ 00:09:53.467 { 00:09:53.467 "dma_device_id": "system", 00:09:53.467 "dma_device_type": 1 00:09:53.467 }, 00:09:53.467 { 00:09:53.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.726 "dma_device_type": 2 00:09:53.726 } 00:09:53.726 ], 00:09:53.726 "driver_specific": {} 00:09:53.726 } 00:09:53.726 ] 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.726 "name": "Existed_Raid", 00:09:53.726 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.726 "strip_size_kb": 64, 00:09:53.726 "state": "configuring", 00:09:53.726 "raid_level": "concat", 00:09:53.726 "superblock": false, 00:09:53.726 "num_base_bdevs": 4, 00:09:53.726 "num_base_bdevs_discovered": 2, 00:09:53.726 "num_base_bdevs_operational": 4, 00:09:53.726 "base_bdevs_list": [ 00:09:53.726 { 00:09:53.726 "name": "BaseBdev1", 00:09:53.726 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:53.726 "is_configured": true, 00:09:53.726 "data_offset": 0, 00:09:53.726 "data_size": 65536 00:09:53.726 }, 00:09:53.726 { 00:09:53.726 "name": "BaseBdev2", 00:09:53.726 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:53.726 "is_configured": true, 00:09:53.726 "data_offset": 0, 00:09:53.726 "data_size": 65536 00:09:53.726 }, 00:09:53.726 { 00:09:53.726 "name": "BaseBdev3", 00:09:53.726 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.726 "is_configured": false, 00:09:53.726 "data_offset": 0, 00:09:53.726 "data_size": 0 00:09:53.726 }, 00:09:53.726 { 00:09:53.726 "name": "BaseBdev4", 00:09:53.726 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.726 "is_configured": false, 00:09:53.726 "data_offset": 0, 00:09:53.726 "data_size": 0 00:09:53.726 } 00:09:53.726 ] 00:09:53.726 }' 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.726 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.986 [2024-09-30 14:08:58.543997] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:53.986 BaseBdev3 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.986 [ 00:09:53.986 { 00:09:53.986 "name": "BaseBdev3", 00:09:53.986 "aliases": [ 00:09:53.986 "3b80962a-1d7d-42ba-828e-f7098695db88" 00:09:53.986 ], 00:09:53.986 "product_name": "Malloc disk", 00:09:53.986 "block_size": 512, 00:09:53.986 "num_blocks": 65536, 00:09:53.986 "uuid": "3b80962a-1d7d-42ba-828e-f7098695db88", 00:09:53.986 "assigned_rate_limits": { 00:09:53.986 "rw_ios_per_sec": 0, 00:09:53.986 "rw_mbytes_per_sec": 0, 00:09:53.986 "r_mbytes_per_sec": 0, 00:09:53.986 "w_mbytes_per_sec": 0 00:09:53.986 }, 00:09:53.986 "claimed": true, 00:09:53.986 "claim_type": "exclusive_write", 00:09:53.986 "zoned": false, 00:09:53.986 "supported_io_types": { 00:09:53.986 "read": true, 00:09:53.986 "write": true, 00:09:53.986 "unmap": true, 00:09:53.986 "flush": true, 00:09:53.986 "reset": true, 00:09:53.986 "nvme_admin": false, 00:09:53.986 "nvme_io": false, 00:09:53.986 "nvme_io_md": false, 00:09:53.986 "write_zeroes": true, 00:09:53.986 "zcopy": true, 00:09:53.986 "get_zone_info": false, 00:09:53.986 "zone_management": false, 00:09:53.986 "zone_append": false, 00:09:53.986 "compare": false, 00:09:53.986 "compare_and_write": false, 00:09:53.986 "abort": true, 00:09:53.986 "seek_hole": false, 00:09:53.986 "seek_data": false, 00:09:53.986 "copy": true, 00:09:53.986 "nvme_iov_md": false 00:09:53.986 }, 00:09:53.986 "memory_domains": [ 00:09:53.986 { 00:09:53.986 "dma_device_id": "system", 00:09:53.986 "dma_device_type": 1 00:09:53.986 }, 00:09:53.986 { 00:09:53.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.986 "dma_device_type": 2 00:09:53.986 } 00:09:53.986 ], 00:09:53.986 "driver_specific": {} 00:09:53.986 } 00:09:53.986 ] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.986 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.986 "name": "Existed_Raid", 00:09:53.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.986 "strip_size_kb": 64, 00:09:53.986 "state": "configuring", 00:09:53.986 "raid_level": "concat", 00:09:53.986 "superblock": false, 00:09:53.986 "num_base_bdevs": 4, 00:09:53.986 "num_base_bdevs_discovered": 3, 00:09:53.986 "num_base_bdevs_operational": 4, 00:09:53.986 "base_bdevs_list": [ 00:09:53.986 { 00:09:53.986 "name": "BaseBdev1", 00:09:53.986 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:53.986 "is_configured": true, 00:09:53.986 "data_offset": 0, 00:09:53.986 "data_size": 65536 00:09:53.986 }, 00:09:53.986 { 00:09:53.986 "name": "BaseBdev2", 00:09:53.986 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:53.986 "is_configured": true, 00:09:53.986 "data_offset": 0, 00:09:53.986 "data_size": 65536 00:09:53.986 }, 00:09:53.986 { 00:09:53.986 "name": "BaseBdev3", 00:09:53.986 "uuid": "3b80962a-1d7d-42ba-828e-f7098695db88", 00:09:53.986 "is_configured": true, 00:09:53.987 "data_offset": 0, 00:09:53.987 "data_size": 65536 00:09:53.987 }, 00:09:53.987 { 00:09:53.987 "name": "BaseBdev4", 00:09:53.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.987 "is_configured": false, 00:09:53.987 "data_offset": 0, 00:09:53.987 "data_size": 0 00:09:53.987 } 00:09:53.987 ] 00:09:53.987 }' 00:09:53.987 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.987 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.554 14:08:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:54.554 14:08:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.554 [2024-09-30 14:08:59.013992] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:54.554 [2024-09-30 14:08:59.014034] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:54.554 [2024-09-30 14:08:59.014043] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:54.554 [2024-09-30 14:08:59.014309] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:54.554 [2024-09-30 14:08:59.014452] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:54.554 [2024-09-30 14:08:59.014462] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:54.554 [2024-09-30 14:08:59.014710] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:54.554 BaseBdev4 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.554 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.554 [ 00:09:54.554 { 00:09:54.554 "name": "BaseBdev4", 00:09:54.554 "aliases": [ 00:09:54.554 "26feaecd-3c56-4162-9227-3f5bcda8c31f" 00:09:54.554 ], 00:09:54.554 "product_name": "Malloc disk", 00:09:54.554 "block_size": 512, 00:09:54.554 "num_blocks": 65536, 00:09:54.554 "uuid": "26feaecd-3c56-4162-9227-3f5bcda8c31f", 00:09:54.554 "assigned_rate_limits": { 00:09:54.554 "rw_ios_per_sec": 0, 00:09:54.554 "rw_mbytes_per_sec": 0, 00:09:54.554 "r_mbytes_per_sec": 0, 00:09:54.554 "w_mbytes_per_sec": 0 00:09:54.554 }, 00:09:54.554 "claimed": true, 00:09:54.555 "claim_type": "exclusive_write", 00:09:54.555 "zoned": false, 00:09:54.555 "supported_io_types": { 00:09:54.555 "read": true, 00:09:54.555 "write": true, 00:09:54.555 "unmap": true, 00:09:54.555 "flush": true, 00:09:54.555 "reset": true, 00:09:54.555 "nvme_admin": false, 00:09:54.555 "nvme_io": false, 00:09:54.555 "nvme_io_md": false, 00:09:54.555 "write_zeroes": true, 00:09:54.555 "zcopy": true, 00:09:54.555 "get_zone_info": false, 00:09:54.555 "zone_management": false, 00:09:54.555 "zone_append": false, 00:09:54.555 "compare": false, 00:09:54.555 "compare_and_write": false, 00:09:54.555 "abort": true, 00:09:54.555 "seek_hole": false, 00:09:54.555 "seek_data": false, 00:09:54.555 "copy": true, 00:09:54.555 "nvme_iov_md": false 00:09:54.555 }, 00:09:54.555 "memory_domains": [ 00:09:54.555 { 00:09:54.555 "dma_device_id": "system", 00:09:54.555 "dma_device_type": 1 00:09:54.555 }, 00:09:54.555 { 00:09:54.555 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.555 "dma_device_type": 2 00:09:54.555 } 00:09:54.555 ], 00:09:54.555 "driver_specific": {} 00:09:54.555 } 00:09:54.555 ] 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.555 "name": "Existed_Raid", 00:09:54.555 "uuid": "778aa72e-cd91-4f9d-b7ca-b1f0cf4b4329", 00:09:54.555 "strip_size_kb": 64, 00:09:54.555 "state": "online", 00:09:54.555 "raid_level": "concat", 00:09:54.555 "superblock": false, 00:09:54.555 "num_base_bdevs": 4, 00:09:54.555 "num_base_bdevs_discovered": 4, 00:09:54.555 "num_base_bdevs_operational": 4, 00:09:54.555 "base_bdevs_list": [ 00:09:54.555 { 00:09:54.555 "name": "BaseBdev1", 00:09:54.555 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:54.555 "is_configured": true, 00:09:54.555 "data_offset": 0, 00:09:54.555 "data_size": 65536 00:09:54.555 }, 00:09:54.555 { 00:09:54.555 "name": "BaseBdev2", 00:09:54.555 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:54.555 "is_configured": true, 00:09:54.555 "data_offset": 0, 00:09:54.555 "data_size": 65536 00:09:54.555 }, 00:09:54.555 { 00:09:54.555 "name": "BaseBdev3", 00:09:54.555 "uuid": "3b80962a-1d7d-42ba-828e-f7098695db88", 00:09:54.555 "is_configured": true, 00:09:54.555 "data_offset": 0, 00:09:54.555 "data_size": 65536 00:09:54.555 }, 00:09:54.555 { 00:09:54.555 "name": "BaseBdev4", 00:09:54.555 "uuid": "26feaecd-3c56-4162-9227-3f5bcda8c31f", 00:09:54.555 "is_configured": true, 00:09:54.555 "data_offset": 0, 00:09:54.555 "data_size": 65536 00:09:54.555 } 00:09:54.555 ] 00:09:54.555 }' 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.555 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:55.124 [2024-09-30 14:08:59.493547] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:55.124 "name": "Existed_Raid", 00:09:55.124 "aliases": [ 00:09:55.124 "778aa72e-cd91-4f9d-b7ca-b1f0cf4b4329" 00:09:55.124 ], 00:09:55.124 "product_name": "Raid Volume", 00:09:55.124 "block_size": 512, 00:09:55.124 "num_blocks": 262144, 00:09:55.124 "uuid": "778aa72e-cd91-4f9d-b7ca-b1f0cf4b4329", 00:09:55.124 "assigned_rate_limits": { 00:09:55.124 "rw_ios_per_sec": 0, 00:09:55.124 "rw_mbytes_per_sec": 0, 00:09:55.124 "r_mbytes_per_sec": 0, 00:09:55.124 "w_mbytes_per_sec": 0 00:09:55.124 }, 00:09:55.124 "claimed": false, 00:09:55.124 "zoned": false, 00:09:55.124 "supported_io_types": { 00:09:55.124 "read": true, 00:09:55.124 "write": true, 00:09:55.124 "unmap": true, 00:09:55.124 "flush": true, 00:09:55.124 "reset": true, 00:09:55.124 "nvme_admin": false, 00:09:55.124 "nvme_io": false, 00:09:55.124 "nvme_io_md": false, 00:09:55.124 "write_zeroes": true, 00:09:55.124 "zcopy": false, 00:09:55.124 "get_zone_info": false, 00:09:55.124 "zone_management": false, 00:09:55.124 "zone_append": false, 00:09:55.124 "compare": false, 00:09:55.124 "compare_and_write": false, 00:09:55.124 "abort": false, 00:09:55.124 "seek_hole": false, 00:09:55.124 "seek_data": false, 00:09:55.124 "copy": false, 00:09:55.124 "nvme_iov_md": false 00:09:55.124 }, 00:09:55.124 "memory_domains": [ 00:09:55.124 { 00:09:55.124 "dma_device_id": "system", 00:09:55.124 "dma_device_type": 1 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.124 "dma_device_type": 2 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "system", 00:09:55.124 "dma_device_type": 1 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.124 "dma_device_type": 2 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "system", 00:09:55.124 "dma_device_type": 1 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.124 "dma_device_type": 2 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "system", 00:09:55.124 "dma_device_type": 1 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.124 "dma_device_type": 2 00:09:55.124 } 00:09:55.124 ], 00:09:55.124 "driver_specific": { 00:09:55.124 "raid": { 00:09:55.124 "uuid": "778aa72e-cd91-4f9d-b7ca-b1f0cf4b4329", 00:09:55.124 "strip_size_kb": 64, 00:09:55.124 "state": "online", 00:09:55.124 "raid_level": "concat", 00:09:55.124 "superblock": false, 00:09:55.124 "num_base_bdevs": 4, 00:09:55.124 "num_base_bdevs_discovered": 4, 00:09:55.124 "num_base_bdevs_operational": 4, 00:09:55.124 "base_bdevs_list": [ 00:09:55.124 { 00:09:55.124 "name": "BaseBdev1", 00:09:55.124 "uuid": "c4141592-4ace-4a12-8013-a97fda723f75", 00:09:55.124 "is_configured": true, 00:09:55.124 "data_offset": 0, 00:09:55.124 "data_size": 65536 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "name": "BaseBdev2", 00:09:55.124 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:55.124 "is_configured": true, 00:09:55.124 "data_offset": 0, 00:09:55.124 "data_size": 65536 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "name": "BaseBdev3", 00:09:55.124 "uuid": "3b80962a-1d7d-42ba-828e-f7098695db88", 00:09:55.124 "is_configured": true, 00:09:55.124 "data_offset": 0, 00:09:55.124 "data_size": 65536 00:09:55.124 }, 00:09:55.124 { 00:09:55.124 "name": "BaseBdev4", 00:09:55.124 "uuid": "26feaecd-3c56-4162-9227-3f5bcda8c31f", 00:09:55.124 "is_configured": true, 00:09:55.124 "data_offset": 0, 00:09:55.124 "data_size": 65536 00:09:55.124 } 00:09:55.124 ] 00:09:55.124 } 00:09:55.124 } 00:09:55.124 }' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:55.124 BaseBdev2 00:09:55.124 BaseBdev3 00:09:55.124 BaseBdev4' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.124 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.125 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.385 [2024-09-30 14:08:59.808690] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:55.385 [2024-09-30 14:08:59.808761] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:55.385 [2024-09-30 14:08:59.808835] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.385 "name": "Existed_Raid", 00:09:55.385 "uuid": "778aa72e-cd91-4f9d-b7ca-b1f0cf4b4329", 00:09:55.385 "strip_size_kb": 64, 00:09:55.385 "state": "offline", 00:09:55.385 "raid_level": "concat", 00:09:55.385 "superblock": false, 00:09:55.385 "num_base_bdevs": 4, 00:09:55.385 "num_base_bdevs_discovered": 3, 00:09:55.385 "num_base_bdevs_operational": 3, 00:09:55.385 "base_bdevs_list": [ 00:09:55.385 { 00:09:55.385 "name": null, 00:09:55.385 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.385 "is_configured": false, 00:09:55.385 "data_offset": 0, 00:09:55.385 "data_size": 65536 00:09:55.385 }, 00:09:55.385 { 00:09:55.385 "name": "BaseBdev2", 00:09:55.385 "uuid": "eaf87b3e-7905-4426-8782-151dccf84103", 00:09:55.385 "is_configured": true, 00:09:55.385 "data_offset": 0, 00:09:55.385 "data_size": 65536 00:09:55.385 }, 00:09:55.385 { 00:09:55.385 "name": "BaseBdev3", 00:09:55.385 "uuid": "3b80962a-1d7d-42ba-828e-f7098695db88", 00:09:55.385 "is_configured": true, 00:09:55.385 "data_offset": 0, 00:09:55.385 "data_size": 65536 00:09:55.385 }, 00:09:55.385 { 00:09:55.385 "name": "BaseBdev4", 00:09:55.385 "uuid": "26feaecd-3c56-4162-9227-3f5bcda8c31f", 00:09:55.385 "is_configured": true, 00:09:55.385 "data_offset": 0, 00:09:55.385 "data_size": 65536 00:09:55.385 } 00:09:55.385 ] 00:09:55.385 }' 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.385 14:08:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.644 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.644 [2024-09-30 14:09:00.295077] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.911 [2024-09-30 14:09:00.366008] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.911 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.911 [2024-09-30 14:09:00.425207] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:55.911 [2024-09-30 14:09:00.425309] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.912 BaseBdev2 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.912 [ 00:09:55.912 { 00:09:55.912 "name": "BaseBdev2", 00:09:55.912 "aliases": [ 00:09:55.912 "b5490fe4-f2b4-455b-abf8-ae97fadb9607" 00:09:55.912 ], 00:09:55.912 "product_name": "Malloc disk", 00:09:55.912 "block_size": 512, 00:09:55.912 "num_blocks": 65536, 00:09:55.912 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:55.912 "assigned_rate_limits": { 00:09:55.912 "rw_ios_per_sec": 0, 00:09:55.912 "rw_mbytes_per_sec": 0, 00:09:55.912 "r_mbytes_per_sec": 0, 00:09:55.912 "w_mbytes_per_sec": 0 00:09:55.912 }, 00:09:55.912 "claimed": false, 00:09:55.912 "zoned": false, 00:09:55.912 "supported_io_types": { 00:09:55.912 "read": true, 00:09:55.912 "write": true, 00:09:55.912 "unmap": true, 00:09:55.912 "flush": true, 00:09:55.912 "reset": true, 00:09:55.912 "nvme_admin": false, 00:09:55.912 "nvme_io": false, 00:09:55.912 "nvme_io_md": false, 00:09:55.912 "write_zeroes": true, 00:09:55.912 "zcopy": true, 00:09:55.912 "get_zone_info": false, 00:09:55.912 "zone_management": false, 00:09:55.912 "zone_append": false, 00:09:55.912 "compare": false, 00:09:55.912 "compare_and_write": false, 00:09:55.912 "abort": true, 00:09:55.912 "seek_hole": false, 00:09:55.912 "seek_data": false, 00:09:55.912 "copy": true, 00:09:55.912 "nvme_iov_md": false 00:09:55.912 }, 00:09:55.912 "memory_domains": [ 00:09:55.912 { 00:09:55.912 "dma_device_id": "system", 00:09:55.912 "dma_device_type": 1 00:09:55.912 }, 00:09:55.912 { 00:09:55.912 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.912 "dma_device_type": 2 00:09:55.912 } 00:09:55.912 ], 00:09:55.912 "driver_specific": {} 00:09:55.912 } 00:09:55.912 ] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.912 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 BaseBdev3 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 [ 00:09:56.198 { 00:09:56.198 "name": "BaseBdev3", 00:09:56.198 "aliases": [ 00:09:56.198 "9cb0e7e2-c201-4444-983b-ef97ad90b90c" 00:09:56.198 ], 00:09:56.198 "product_name": "Malloc disk", 00:09:56.198 "block_size": 512, 00:09:56.198 "num_blocks": 65536, 00:09:56.198 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:56.198 "assigned_rate_limits": { 00:09:56.198 "rw_ios_per_sec": 0, 00:09:56.198 "rw_mbytes_per_sec": 0, 00:09:56.198 "r_mbytes_per_sec": 0, 00:09:56.198 "w_mbytes_per_sec": 0 00:09:56.198 }, 00:09:56.198 "claimed": false, 00:09:56.198 "zoned": false, 00:09:56.198 "supported_io_types": { 00:09:56.198 "read": true, 00:09:56.198 "write": true, 00:09:56.198 "unmap": true, 00:09:56.198 "flush": true, 00:09:56.198 "reset": true, 00:09:56.198 "nvme_admin": false, 00:09:56.198 "nvme_io": false, 00:09:56.198 "nvme_io_md": false, 00:09:56.198 "write_zeroes": true, 00:09:56.198 "zcopy": true, 00:09:56.198 "get_zone_info": false, 00:09:56.198 "zone_management": false, 00:09:56.198 "zone_append": false, 00:09:56.198 "compare": false, 00:09:56.198 "compare_and_write": false, 00:09:56.198 "abort": true, 00:09:56.198 "seek_hole": false, 00:09:56.198 "seek_data": false, 00:09:56.198 "copy": true, 00:09:56.198 "nvme_iov_md": false 00:09:56.198 }, 00:09:56.198 "memory_domains": [ 00:09:56.198 { 00:09:56.198 "dma_device_id": "system", 00:09:56.198 "dma_device_type": 1 00:09:56.198 }, 00:09:56.198 { 00:09:56.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.198 "dma_device_type": 2 00:09:56.198 } 00:09:56.198 ], 00:09:56.198 "driver_specific": {} 00:09:56.198 } 00:09:56.198 ] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 BaseBdev4 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.198 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.198 [ 00:09:56.198 { 00:09:56.198 "name": "BaseBdev4", 00:09:56.198 "aliases": [ 00:09:56.198 "683ce911-2d17-46aa-af3f-f141e5c0f41b" 00:09:56.198 ], 00:09:56.198 "product_name": "Malloc disk", 00:09:56.198 "block_size": 512, 00:09:56.198 "num_blocks": 65536, 00:09:56.198 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:56.198 "assigned_rate_limits": { 00:09:56.198 "rw_ios_per_sec": 0, 00:09:56.198 "rw_mbytes_per_sec": 0, 00:09:56.198 "r_mbytes_per_sec": 0, 00:09:56.198 "w_mbytes_per_sec": 0 00:09:56.198 }, 00:09:56.198 "claimed": false, 00:09:56.198 "zoned": false, 00:09:56.198 "supported_io_types": { 00:09:56.198 "read": true, 00:09:56.198 "write": true, 00:09:56.198 "unmap": true, 00:09:56.198 "flush": true, 00:09:56.198 "reset": true, 00:09:56.198 "nvme_admin": false, 00:09:56.198 "nvme_io": false, 00:09:56.198 "nvme_io_md": false, 00:09:56.198 "write_zeroes": true, 00:09:56.198 "zcopy": true, 00:09:56.198 "get_zone_info": false, 00:09:56.198 "zone_management": false, 00:09:56.199 "zone_append": false, 00:09:56.199 "compare": false, 00:09:56.199 "compare_and_write": false, 00:09:56.199 "abort": true, 00:09:56.199 "seek_hole": false, 00:09:56.199 "seek_data": false, 00:09:56.199 "copy": true, 00:09:56.199 "nvme_iov_md": false 00:09:56.199 }, 00:09:56.199 "memory_domains": [ 00:09:56.199 { 00:09:56.199 "dma_device_id": "system", 00:09:56.199 "dma_device_type": 1 00:09:56.199 }, 00:09:56.199 { 00:09:56.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.199 "dma_device_type": 2 00:09:56.199 } 00:09:56.199 ], 00:09:56.199 "driver_specific": {} 00:09:56.199 } 00:09:56.199 ] 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.199 [2024-09-30 14:09:00.656531] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:56.199 [2024-09-30 14:09:00.656613] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:56.199 [2024-09-30 14:09:00.656649] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:56.199 [2024-09-30 14:09:00.658413] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:56.199 [2024-09-30 14:09:00.658502] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.199 "name": "Existed_Raid", 00:09:56.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.199 "strip_size_kb": 64, 00:09:56.199 "state": "configuring", 00:09:56.199 "raid_level": "concat", 00:09:56.199 "superblock": false, 00:09:56.199 "num_base_bdevs": 4, 00:09:56.199 "num_base_bdevs_discovered": 3, 00:09:56.199 "num_base_bdevs_operational": 4, 00:09:56.199 "base_bdevs_list": [ 00:09:56.199 { 00:09:56.199 "name": "BaseBdev1", 00:09:56.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.199 "is_configured": false, 00:09:56.199 "data_offset": 0, 00:09:56.199 "data_size": 0 00:09:56.199 }, 00:09:56.199 { 00:09:56.199 "name": "BaseBdev2", 00:09:56.199 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:56.199 "is_configured": true, 00:09:56.199 "data_offset": 0, 00:09:56.199 "data_size": 65536 00:09:56.199 }, 00:09:56.199 { 00:09:56.199 "name": "BaseBdev3", 00:09:56.199 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:56.199 "is_configured": true, 00:09:56.199 "data_offset": 0, 00:09:56.199 "data_size": 65536 00:09:56.199 }, 00:09:56.199 { 00:09:56.199 "name": "BaseBdev4", 00:09:56.199 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:56.199 "is_configured": true, 00:09:56.199 "data_offset": 0, 00:09:56.199 "data_size": 65536 00:09:56.199 } 00:09:56.199 ] 00:09:56.199 }' 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.199 14:09:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.475 [2024-09-30 14:09:01.059829] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.475 "name": "Existed_Raid", 00:09:56.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.475 "strip_size_kb": 64, 00:09:56.475 "state": "configuring", 00:09:56.475 "raid_level": "concat", 00:09:56.475 "superblock": false, 00:09:56.475 "num_base_bdevs": 4, 00:09:56.475 "num_base_bdevs_discovered": 2, 00:09:56.475 "num_base_bdevs_operational": 4, 00:09:56.475 "base_bdevs_list": [ 00:09:56.475 { 00:09:56.475 "name": "BaseBdev1", 00:09:56.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.475 "is_configured": false, 00:09:56.475 "data_offset": 0, 00:09:56.475 "data_size": 0 00:09:56.475 }, 00:09:56.475 { 00:09:56.475 "name": null, 00:09:56.475 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:56.475 "is_configured": false, 00:09:56.475 "data_offset": 0, 00:09:56.475 "data_size": 65536 00:09:56.475 }, 00:09:56.475 { 00:09:56.475 "name": "BaseBdev3", 00:09:56.475 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:56.475 "is_configured": true, 00:09:56.475 "data_offset": 0, 00:09:56.475 "data_size": 65536 00:09:56.475 }, 00:09:56.475 { 00:09:56.475 "name": "BaseBdev4", 00:09:56.475 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:56.475 "is_configured": true, 00:09:56.475 "data_offset": 0, 00:09:56.475 "data_size": 65536 00:09:56.475 } 00:09:56.475 ] 00:09:56.475 }' 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.475 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 [2024-09-30 14:09:01.537671] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:57.044 BaseBdev1 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 [ 00:09:57.044 { 00:09:57.044 "name": "BaseBdev1", 00:09:57.044 "aliases": [ 00:09:57.044 "576ebb6e-2732-437a-acdd-b5b78dd727f6" 00:09:57.044 ], 00:09:57.044 "product_name": "Malloc disk", 00:09:57.044 "block_size": 512, 00:09:57.044 "num_blocks": 65536, 00:09:57.044 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:57.044 "assigned_rate_limits": { 00:09:57.044 "rw_ios_per_sec": 0, 00:09:57.044 "rw_mbytes_per_sec": 0, 00:09:57.044 "r_mbytes_per_sec": 0, 00:09:57.044 "w_mbytes_per_sec": 0 00:09:57.044 }, 00:09:57.044 "claimed": true, 00:09:57.044 "claim_type": "exclusive_write", 00:09:57.044 "zoned": false, 00:09:57.044 "supported_io_types": { 00:09:57.044 "read": true, 00:09:57.044 "write": true, 00:09:57.044 "unmap": true, 00:09:57.044 "flush": true, 00:09:57.044 "reset": true, 00:09:57.044 "nvme_admin": false, 00:09:57.044 "nvme_io": false, 00:09:57.044 "nvme_io_md": false, 00:09:57.044 "write_zeroes": true, 00:09:57.044 "zcopy": true, 00:09:57.044 "get_zone_info": false, 00:09:57.044 "zone_management": false, 00:09:57.044 "zone_append": false, 00:09:57.044 "compare": false, 00:09:57.044 "compare_and_write": false, 00:09:57.044 "abort": true, 00:09:57.044 "seek_hole": false, 00:09:57.044 "seek_data": false, 00:09:57.044 "copy": true, 00:09:57.044 "nvme_iov_md": false 00:09:57.044 }, 00:09:57.044 "memory_domains": [ 00:09:57.044 { 00:09:57.044 "dma_device_id": "system", 00:09:57.044 "dma_device_type": 1 00:09:57.044 }, 00:09:57.044 { 00:09:57.044 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.044 "dma_device_type": 2 00:09:57.044 } 00:09:57.044 ], 00:09:57.044 "driver_specific": {} 00:09:57.044 } 00:09:57.044 ] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.044 "name": "Existed_Raid", 00:09:57.044 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.044 "strip_size_kb": 64, 00:09:57.044 "state": "configuring", 00:09:57.044 "raid_level": "concat", 00:09:57.044 "superblock": false, 00:09:57.044 "num_base_bdevs": 4, 00:09:57.044 "num_base_bdevs_discovered": 3, 00:09:57.044 "num_base_bdevs_operational": 4, 00:09:57.044 "base_bdevs_list": [ 00:09:57.044 { 00:09:57.044 "name": "BaseBdev1", 00:09:57.044 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:57.044 "is_configured": true, 00:09:57.044 "data_offset": 0, 00:09:57.044 "data_size": 65536 00:09:57.044 }, 00:09:57.044 { 00:09:57.044 "name": null, 00:09:57.044 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:57.044 "is_configured": false, 00:09:57.044 "data_offset": 0, 00:09:57.044 "data_size": 65536 00:09:57.044 }, 00:09:57.044 { 00:09:57.044 "name": "BaseBdev3", 00:09:57.044 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:57.044 "is_configured": true, 00:09:57.044 "data_offset": 0, 00:09:57.044 "data_size": 65536 00:09:57.044 }, 00:09:57.044 { 00:09:57.044 "name": "BaseBdev4", 00:09:57.044 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:57.044 "is_configured": true, 00:09:57.044 "data_offset": 0, 00:09:57.044 "data_size": 65536 00:09:57.044 } 00:09:57.044 ] 00:09:57.044 }' 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.044 14:09:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.612 [2024-09-30 14:09:02.068817] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.612 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.613 "name": "Existed_Raid", 00:09:57.613 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.613 "strip_size_kb": 64, 00:09:57.613 "state": "configuring", 00:09:57.613 "raid_level": "concat", 00:09:57.613 "superblock": false, 00:09:57.613 "num_base_bdevs": 4, 00:09:57.613 "num_base_bdevs_discovered": 2, 00:09:57.613 "num_base_bdevs_operational": 4, 00:09:57.613 "base_bdevs_list": [ 00:09:57.613 { 00:09:57.613 "name": "BaseBdev1", 00:09:57.613 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:57.613 "is_configured": true, 00:09:57.613 "data_offset": 0, 00:09:57.613 "data_size": 65536 00:09:57.613 }, 00:09:57.613 { 00:09:57.613 "name": null, 00:09:57.613 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:57.613 "is_configured": false, 00:09:57.613 "data_offset": 0, 00:09:57.613 "data_size": 65536 00:09:57.613 }, 00:09:57.613 { 00:09:57.613 "name": null, 00:09:57.613 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:57.613 "is_configured": false, 00:09:57.613 "data_offset": 0, 00:09:57.613 "data_size": 65536 00:09:57.613 }, 00:09:57.613 { 00:09:57.613 "name": "BaseBdev4", 00:09:57.613 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:57.613 "is_configured": true, 00:09:57.613 "data_offset": 0, 00:09:57.613 "data_size": 65536 00:09:57.613 } 00:09:57.613 ] 00:09:57.613 }' 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.613 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 [2024-09-30 14:09:02.599933] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.181 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.181 "name": "Existed_Raid", 00:09:58.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.181 "strip_size_kb": 64, 00:09:58.181 "state": "configuring", 00:09:58.181 "raid_level": "concat", 00:09:58.181 "superblock": false, 00:09:58.181 "num_base_bdevs": 4, 00:09:58.181 "num_base_bdevs_discovered": 3, 00:09:58.181 "num_base_bdevs_operational": 4, 00:09:58.181 "base_bdevs_list": [ 00:09:58.181 { 00:09:58.181 "name": "BaseBdev1", 00:09:58.181 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:58.182 "is_configured": true, 00:09:58.182 "data_offset": 0, 00:09:58.182 "data_size": 65536 00:09:58.182 }, 00:09:58.182 { 00:09:58.182 "name": null, 00:09:58.182 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:58.182 "is_configured": false, 00:09:58.182 "data_offset": 0, 00:09:58.182 "data_size": 65536 00:09:58.182 }, 00:09:58.182 { 00:09:58.182 "name": "BaseBdev3", 00:09:58.182 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:58.182 "is_configured": true, 00:09:58.182 "data_offset": 0, 00:09:58.182 "data_size": 65536 00:09:58.182 }, 00:09:58.182 { 00:09:58.182 "name": "BaseBdev4", 00:09:58.182 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:58.182 "is_configured": true, 00:09:58.182 "data_offset": 0, 00:09:58.182 "data_size": 65536 00:09:58.182 } 00:09:58.182 ] 00:09:58.182 }' 00:09:58.182 14:09:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.182 14:09:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.442 [2024-09-30 14:09:03.075116] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.442 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.702 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.702 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.702 "name": "Existed_Raid", 00:09:58.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.702 "strip_size_kb": 64, 00:09:58.702 "state": "configuring", 00:09:58.702 "raid_level": "concat", 00:09:58.702 "superblock": false, 00:09:58.702 "num_base_bdevs": 4, 00:09:58.702 "num_base_bdevs_discovered": 2, 00:09:58.702 "num_base_bdevs_operational": 4, 00:09:58.702 "base_bdevs_list": [ 00:09:58.702 { 00:09:58.702 "name": null, 00:09:58.702 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:58.702 "is_configured": false, 00:09:58.702 "data_offset": 0, 00:09:58.702 "data_size": 65536 00:09:58.702 }, 00:09:58.702 { 00:09:58.702 "name": null, 00:09:58.702 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:58.702 "is_configured": false, 00:09:58.702 "data_offset": 0, 00:09:58.702 "data_size": 65536 00:09:58.702 }, 00:09:58.702 { 00:09:58.702 "name": "BaseBdev3", 00:09:58.702 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:58.702 "is_configured": true, 00:09:58.702 "data_offset": 0, 00:09:58.702 "data_size": 65536 00:09:58.702 }, 00:09:58.702 { 00:09:58.702 "name": "BaseBdev4", 00:09:58.702 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:58.702 "is_configured": true, 00:09:58.702 "data_offset": 0, 00:09:58.702 "data_size": 65536 00:09:58.702 } 00:09:58.702 ] 00:09:58.702 }' 00:09:58.702 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.702 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.961 [2024-09-30 14:09:03.540624] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.961 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.961 "name": "Existed_Raid", 00:09:58.961 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.961 "strip_size_kb": 64, 00:09:58.961 "state": "configuring", 00:09:58.961 "raid_level": "concat", 00:09:58.961 "superblock": false, 00:09:58.961 "num_base_bdevs": 4, 00:09:58.961 "num_base_bdevs_discovered": 3, 00:09:58.961 "num_base_bdevs_operational": 4, 00:09:58.961 "base_bdevs_list": [ 00:09:58.961 { 00:09:58.961 "name": null, 00:09:58.961 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:58.961 "is_configured": false, 00:09:58.961 "data_offset": 0, 00:09:58.961 "data_size": 65536 00:09:58.961 }, 00:09:58.961 { 00:09:58.961 "name": "BaseBdev2", 00:09:58.961 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:58.961 "is_configured": true, 00:09:58.961 "data_offset": 0, 00:09:58.961 "data_size": 65536 00:09:58.961 }, 00:09:58.961 { 00:09:58.961 "name": "BaseBdev3", 00:09:58.961 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:58.961 "is_configured": true, 00:09:58.961 "data_offset": 0, 00:09:58.961 "data_size": 65536 00:09:58.961 }, 00:09:58.961 { 00:09:58.961 "name": "BaseBdev4", 00:09:58.961 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:58.961 "is_configured": true, 00:09:58.962 "data_offset": 0, 00:09:58.962 "data_size": 65536 00:09:58.962 } 00:09:58.962 ] 00:09:58.962 }' 00:09:58.962 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.962 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:59.531 14:09:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 576ebb6e-2732-437a-acdd-b5b78dd727f6 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 [2024-09-30 14:09:04.034554] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:59.531 [2024-09-30 14:09:04.034648] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:59.531 [2024-09-30 14:09:04.034674] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:59.531 [2024-09-30 14:09:04.034966] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:09:59.531 [2024-09-30 14:09:04.035110] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:59.531 [2024-09-30 14:09:04.035146] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:59.531 [2024-09-30 14:09:04.035352] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:59.531 NewBaseBdev 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.531 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.531 [ 00:09:59.531 { 00:09:59.531 "name": "NewBaseBdev", 00:09:59.531 "aliases": [ 00:09:59.531 "576ebb6e-2732-437a-acdd-b5b78dd727f6" 00:09:59.531 ], 00:09:59.531 "product_name": "Malloc disk", 00:09:59.531 "block_size": 512, 00:09:59.531 "num_blocks": 65536, 00:09:59.531 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:59.531 "assigned_rate_limits": { 00:09:59.531 "rw_ios_per_sec": 0, 00:09:59.531 "rw_mbytes_per_sec": 0, 00:09:59.531 "r_mbytes_per_sec": 0, 00:09:59.531 "w_mbytes_per_sec": 0 00:09:59.531 }, 00:09:59.531 "claimed": true, 00:09:59.531 "claim_type": "exclusive_write", 00:09:59.531 "zoned": false, 00:09:59.531 "supported_io_types": { 00:09:59.531 "read": true, 00:09:59.531 "write": true, 00:09:59.531 "unmap": true, 00:09:59.531 "flush": true, 00:09:59.531 "reset": true, 00:09:59.531 "nvme_admin": false, 00:09:59.531 "nvme_io": false, 00:09:59.531 "nvme_io_md": false, 00:09:59.531 "write_zeroes": true, 00:09:59.531 "zcopy": true, 00:09:59.531 "get_zone_info": false, 00:09:59.532 "zone_management": false, 00:09:59.532 "zone_append": false, 00:09:59.532 "compare": false, 00:09:59.532 "compare_and_write": false, 00:09:59.532 "abort": true, 00:09:59.532 "seek_hole": false, 00:09:59.532 "seek_data": false, 00:09:59.532 "copy": true, 00:09:59.532 "nvme_iov_md": false 00:09:59.532 }, 00:09:59.532 "memory_domains": [ 00:09:59.532 { 00:09:59.532 "dma_device_id": "system", 00:09:59.532 "dma_device_type": 1 00:09:59.532 }, 00:09:59.532 { 00:09:59.532 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.532 "dma_device_type": 2 00:09:59.532 } 00:09:59.532 ], 00:09:59.532 "driver_specific": {} 00:09:59.532 } 00:09:59.532 ] 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.532 "name": "Existed_Raid", 00:09:59.532 "uuid": "4b1aa555-c077-46d9-a5aa-80e4eb0ce3c7", 00:09:59.532 "strip_size_kb": 64, 00:09:59.532 "state": "online", 00:09:59.532 "raid_level": "concat", 00:09:59.532 "superblock": false, 00:09:59.532 "num_base_bdevs": 4, 00:09:59.532 "num_base_bdevs_discovered": 4, 00:09:59.532 "num_base_bdevs_operational": 4, 00:09:59.532 "base_bdevs_list": [ 00:09:59.532 { 00:09:59.532 "name": "NewBaseBdev", 00:09:59.532 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:09:59.532 "is_configured": true, 00:09:59.532 "data_offset": 0, 00:09:59.532 "data_size": 65536 00:09:59.532 }, 00:09:59.532 { 00:09:59.532 "name": "BaseBdev2", 00:09:59.532 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:09:59.532 "is_configured": true, 00:09:59.532 "data_offset": 0, 00:09:59.532 "data_size": 65536 00:09:59.532 }, 00:09:59.532 { 00:09:59.532 "name": "BaseBdev3", 00:09:59.532 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:09:59.532 "is_configured": true, 00:09:59.532 "data_offset": 0, 00:09:59.532 "data_size": 65536 00:09:59.532 }, 00:09:59.532 { 00:09:59.532 "name": "BaseBdev4", 00:09:59.532 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:09:59.532 "is_configured": true, 00:09:59.532 "data_offset": 0, 00:09:59.532 "data_size": 65536 00:09:59.532 } 00:09:59.532 ] 00:09:59.532 }' 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.532 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.792 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:59.792 [2024-09-30 14:09:04.446224] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:00.052 "name": "Existed_Raid", 00:10:00.052 "aliases": [ 00:10:00.052 "4b1aa555-c077-46d9-a5aa-80e4eb0ce3c7" 00:10:00.052 ], 00:10:00.052 "product_name": "Raid Volume", 00:10:00.052 "block_size": 512, 00:10:00.052 "num_blocks": 262144, 00:10:00.052 "uuid": "4b1aa555-c077-46d9-a5aa-80e4eb0ce3c7", 00:10:00.052 "assigned_rate_limits": { 00:10:00.052 "rw_ios_per_sec": 0, 00:10:00.052 "rw_mbytes_per_sec": 0, 00:10:00.052 "r_mbytes_per_sec": 0, 00:10:00.052 "w_mbytes_per_sec": 0 00:10:00.052 }, 00:10:00.052 "claimed": false, 00:10:00.052 "zoned": false, 00:10:00.052 "supported_io_types": { 00:10:00.052 "read": true, 00:10:00.052 "write": true, 00:10:00.052 "unmap": true, 00:10:00.052 "flush": true, 00:10:00.052 "reset": true, 00:10:00.052 "nvme_admin": false, 00:10:00.052 "nvme_io": false, 00:10:00.052 "nvme_io_md": false, 00:10:00.052 "write_zeroes": true, 00:10:00.052 "zcopy": false, 00:10:00.052 "get_zone_info": false, 00:10:00.052 "zone_management": false, 00:10:00.052 "zone_append": false, 00:10:00.052 "compare": false, 00:10:00.052 "compare_and_write": false, 00:10:00.052 "abort": false, 00:10:00.052 "seek_hole": false, 00:10:00.052 "seek_data": false, 00:10:00.052 "copy": false, 00:10:00.052 "nvme_iov_md": false 00:10:00.052 }, 00:10:00.052 "memory_domains": [ 00:10:00.052 { 00:10:00.052 "dma_device_id": "system", 00:10:00.052 "dma_device_type": 1 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.052 "dma_device_type": 2 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "system", 00:10:00.052 "dma_device_type": 1 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.052 "dma_device_type": 2 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "system", 00:10:00.052 "dma_device_type": 1 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.052 "dma_device_type": 2 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "system", 00:10:00.052 "dma_device_type": 1 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.052 "dma_device_type": 2 00:10:00.052 } 00:10:00.052 ], 00:10:00.052 "driver_specific": { 00:10:00.052 "raid": { 00:10:00.052 "uuid": "4b1aa555-c077-46d9-a5aa-80e4eb0ce3c7", 00:10:00.052 "strip_size_kb": 64, 00:10:00.052 "state": "online", 00:10:00.052 "raid_level": "concat", 00:10:00.052 "superblock": false, 00:10:00.052 "num_base_bdevs": 4, 00:10:00.052 "num_base_bdevs_discovered": 4, 00:10:00.052 "num_base_bdevs_operational": 4, 00:10:00.052 "base_bdevs_list": [ 00:10:00.052 { 00:10:00.052 "name": "NewBaseBdev", 00:10:00.052 "uuid": "576ebb6e-2732-437a-acdd-b5b78dd727f6", 00:10:00.052 "is_configured": true, 00:10:00.052 "data_offset": 0, 00:10:00.052 "data_size": 65536 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "name": "BaseBdev2", 00:10:00.052 "uuid": "b5490fe4-f2b4-455b-abf8-ae97fadb9607", 00:10:00.052 "is_configured": true, 00:10:00.052 "data_offset": 0, 00:10:00.052 "data_size": 65536 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "name": "BaseBdev3", 00:10:00.052 "uuid": "9cb0e7e2-c201-4444-983b-ef97ad90b90c", 00:10:00.052 "is_configured": true, 00:10:00.052 "data_offset": 0, 00:10:00.052 "data_size": 65536 00:10:00.052 }, 00:10:00.052 { 00:10:00.052 "name": "BaseBdev4", 00:10:00.052 "uuid": "683ce911-2d17-46aa-af3f-f141e5c0f41b", 00:10:00.052 "is_configured": true, 00:10:00.052 "data_offset": 0, 00:10:00.052 "data_size": 65536 00:10:00.052 } 00:10:00.052 ] 00:10:00.052 } 00:10:00.052 } 00:10:00.052 }' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:00.052 BaseBdev2 00:10:00.052 BaseBdev3 00:10:00.052 BaseBdev4' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.052 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.311 [2024-09-30 14:09:04.749342] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:00.311 [2024-09-30 14:09:04.749369] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:00.311 [2024-09-30 14:09:04.749449] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:00.311 [2024-09-30 14:09:04.749524] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:00.311 [2024-09-30 14:09:04.749545] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 83201 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 83201 ']' 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 83201 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83201 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:00.311 killing process with pid 83201 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83201' 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 83201 00:10:00.311 [2024-09-30 14:09:04.792791] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:00.311 14:09:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 83201 00:10:00.311 [2024-09-30 14:09:04.832515] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:00.570 00:10:00.570 real 0m9.273s 00:10:00.570 user 0m15.729s 00:10:00.570 sys 0m1.987s 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.570 ************************************ 00:10:00.570 END TEST raid_state_function_test 00:10:00.570 ************************************ 00:10:00.570 14:09:05 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:10:00.570 14:09:05 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:00.570 14:09:05 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:00.570 14:09:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:00.570 ************************************ 00:10:00.570 START TEST raid_state_function_test_sb 00:10:00.570 ************************************ 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 true 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:00.570 Process raid pid: 83845 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=83845 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83845' 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 83845 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 83845 ']' 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:00.570 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:00.570 14:09:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.829 [2024-09-30 14:09:05.250776] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:00.829 [2024-09-30 14:09:05.250907] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:00.830 [2024-09-30 14:09:05.383738] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:00.830 [2024-09-30 14:09:05.412980] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:00.830 [2024-09-30 14:09:05.458856] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:01.098 [2024-09-30 14:09:05.500269] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:01.098 [2024-09-30 14:09:05.500388] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.668 [2024-09-30 14:09:06.073267] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:01.668 [2024-09-30 14:09:06.073319] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:01.668 [2024-09-30 14:09:06.073331] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:01.668 [2024-09-30 14:09:06.073339] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:01.668 [2024-09-30 14:09:06.073350] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:01.668 [2024-09-30 14:09:06.073357] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:01.668 [2024-09-30 14:09:06.073364] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:01.668 [2024-09-30 14:09:06.073371] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.668 "name": "Existed_Raid", 00:10:01.668 "uuid": "9d8e16f3-964f-4488-bec0-3404782dee4e", 00:10:01.668 "strip_size_kb": 64, 00:10:01.668 "state": "configuring", 00:10:01.668 "raid_level": "concat", 00:10:01.668 "superblock": true, 00:10:01.668 "num_base_bdevs": 4, 00:10:01.668 "num_base_bdevs_discovered": 0, 00:10:01.668 "num_base_bdevs_operational": 4, 00:10:01.668 "base_bdevs_list": [ 00:10:01.668 { 00:10:01.668 "name": "BaseBdev1", 00:10:01.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.668 "is_configured": false, 00:10:01.668 "data_offset": 0, 00:10:01.668 "data_size": 0 00:10:01.668 }, 00:10:01.668 { 00:10:01.668 "name": "BaseBdev2", 00:10:01.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.668 "is_configured": false, 00:10:01.668 "data_offset": 0, 00:10:01.668 "data_size": 0 00:10:01.668 }, 00:10:01.668 { 00:10:01.668 "name": "BaseBdev3", 00:10:01.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.668 "is_configured": false, 00:10:01.668 "data_offset": 0, 00:10:01.668 "data_size": 0 00:10:01.668 }, 00:10:01.668 { 00:10:01.668 "name": "BaseBdev4", 00:10:01.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.668 "is_configured": false, 00:10:01.668 "data_offset": 0, 00:10:01.668 "data_size": 0 00:10:01.668 } 00:10:01.668 ] 00:10:01.668 }' 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.668 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.927 [2024-09-30 14:09:06.560309] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:01.927 [2024-09-30 14:09:06.560385] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.927 [2024-09-30 14:09:06.572312] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:01.927 [2024-09-30 14:09:06.572381] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:01.927 [2024-09-30 14:09:06.572411] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:01.927 [2024-09-30 14:09:06.572431] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:01.927 [2024-09-30 14:09:06.572463] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:01.927 [2024-09-30 14:09:06.572512] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:01.927 [2024-09-30 14:09:06.572545] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:01.927 [2024-09-30 14:09:06.572567] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.927 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.187 [2024-09-30 14:09:06.593092] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:02.187 BaseBdev1 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.187 [ 00:10:02.187 { 00:10:02.187 "name": "BaseBdev1", 00:10:02.187 "aliases": [ 00:10:02.187 "a7413b80-74d1-4d06-ba0b-9aa77572d575" 00:10:02.187 ], 00:10:02.187 "product_name": "Malloc disk", 00:10:02.187 "block_size": 512, 00:10:02.187 "num_blocks": 65536, 00:10:02.187 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:02.187 "assigned_rate_limits": { 00:10:02.187 "rw_ios_per_sec": 0, 00:10:02.187 "rw_mbytes_per_sec": 0, 00:10:02.187 "r_mbytes_per_sec": 0, 00:10:02.187 "w_mbytes_per_sec": 0 00:10:02.187 }, 00:10:02.187 "claimed": true, 00:10:02.187 "claim_type": "exclusive_write", 00:10:02.187 "zoned": false, 00:10:02.187 "supported_io_types": { 00:10:02.187 "read": true, 00:10:02.187 "write": true, 00:10:02.187 "unmap": true, 00:10:02.187 "flush": true, 00:10:02.187 "reset": true, 00:10:02.187 "nvme_admin": false, 00:10:02.187 "nvme_io": false, 00:10:02.187 "nvme_io_md": false, 00:10:02.187 "write_zeroes": true, 00:10:02.187 "zcopy": true, 00:10:02.187 "get_zone_info": false, 00:10:02.187 "zone_management": false, 00:10:02.187 "zone_append": false, 00:10:02.187 "compare": false, 00:10:02.187 "compare_and_write": false, 00:10:02.187 "abort": true, 00:10:02.187 "seek_hole": false, 00:10:02.187 "seek_data": false, 00:10:02.187 "copy": true, 00:10:02.187 "nvme_iov_md": false 00:10:02.187 }, 00:10:02.187 "memory_domains": [ 00:10:02.187 { 00:10:02.187 "dma_device_id": "system", 00:10:02.187 "dma_device_type": 1 00:10:02.187 }, 00:10:02.187 { 00:10:02.187 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.187 "dma_device_type": 2 00:10:02.187 } 00:10:02.187 ], 00:10:02.187 "driver_specific": {} 00:10:02.187 } 00:10:02.187 ] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.187 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.187 "name": "Existed_Raid", 00:10:02.187 "uuid": "32548dda-9c9e-480e-83c3-fdd5da462ed2", 00:10:02.187 "strip_size_kb": 64, 00:10:02.187 "state": "configuring", 00:10:02.187 "raid_level": "concat", 00:10:02.187 "superblock": true, 00:10:02.187 "num_base_bdevs": 4, 00:10:02.187 "num_base_bdevs_discovered": 1, 00:10:02.187 "num_base_bdevs_operational": 4, 00:10:02.187 "base_bdevs_list": [ 00:10:02.187 { 00:10:02.187 "name": "BaseBdev1", 00:10:02.187 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:02.187 "is_configured": true, 00:10:02.187 "data_offset": 2048, 00:10:02.187 "data_size": 63488 00:10:02.187 }, 00:10:02.187 { 00:10:02.187 "name": "BaseBdev2", 00:10:02.187 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.188 "is_configured": false, 00:10:02.188 "data_offset": 0, 00:10:02.188 "data_size": 0 00:10:02.188 }, 00:10:02.188 { 00:10:02.188 "name": "BaseBdev3", 00:10:02.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.188 "is_configured": false, 00:10:02.188 "data_offset": 0, 00:10:02.188 "data_size": 0 00:10:02.188 }, 00:10:02.188 { 00:10:02.188 "name": "BaseBdev4", 00:10:02.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.188 "is_configured": false, 00:10:02.188 "data_offset": 0, 00:10:02.188 "data_size": 0 00:10:02.188 } 00:10:02.188 ] 00:10:02.188 }' 00:10:02.188 14:09:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.188 14:09:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.446 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:02.446 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.447 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.705 [2024-09-30 14:09:07.108226] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:02.705 [2024-09-30 14:09:07.108328] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.705 [2024-09-30 14:09:07.120219] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:02.705 [2024-09-30 14:09:07.121979] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:02.705 [2024-09-30 14:09:07.122045] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:02.705 [2024-09-30 14:09:07.122089] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:02.705 [2024-09-30 14:09:07.122109] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:02.705 [2024-09-30 14:09:07.122127] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:02.705 [2024-09-30 14:09:07.122145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.705 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.706 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.706 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.706 "name": "Existed_Raid", 00:10:02.706 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:02.706 "strip_size_kb": 64, 00:10:02.706 "state": "configuring", 00:10:02.706 "raid_level": "concat", 00:10:02.706 "superblock": true, 00:10:02.706 "num_base_bdevs": 4, 00:10:02.706 "num_base_bdevs_discovered": 1, 00:10:02.706 "num_base_bdevs_operational": 4, 00:10:02.706 "base_bdevs_list": [ 00:10:02.706 { 00:10:02.706 "name": "BaseBdev1", 00:10:02.706 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:02.706 "is_configured": true, 00:10:02.706 "data_offset": 2048, 00:10:02.706 "data_size": 63488 00:10:02.706 }, 00:10:02.706 { 00:10:02.706 "name": "BaseBdev2", 00:10:02.706 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.706 "is_configured": false, 00:10:02.706 "data_offset": 0, 00:10:02.706 "data_size": 0 00:10:02.706 }, 00:10:02.706 { 00:10:02.706 "name": "BaseBdev3", 00:10:02.706 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.706 "is_configured": false, 00:10:02.706 "data_offset": 0, 00:10:02.706 "data_size": 0 00:10:02.706 }, 00:10:02.706 { 00:10:02.706 "name": "BaseBdev4", 00:10:02.706 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.706 "is_configured": false, 00:10:02.706 "data_offset": 0, 00:10:02.706 "data_size": 0 00:10:02.706 } 00:10:02.706 ] 00:10:02.706 }' 00:10:02.706 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.706 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.965 [2024-09-30 14:09:07.576345] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:02.965 BaseBdev2 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.965 [ 00:10:02.965 { 00:10:02.965 "name": "BaseBdev2", 00:10:02.965 "aliases": [ 00:10:02.965 "4e1e3775-68a4-4275-b197-4a233258ceb5" 00:10:02.965 ], 00:10:02.965 "product_name": "Malloc disk", 00:10:02.965 "block_size": 512, 00:10:02.965 "num_blocks": 65536, 00:10:02.965 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:02.965 "assigned_rate_limits": { 00:10:02.965 "rw_ios_per_sec": 0, 00:10:02.965 "rw_mbytes_per_sec": 0, 00:10:02.965 "r_mbytes_per_sec": 0, 00:10:02.965 "w_mbytes_per_sec": 0 00:10:02.965 }, 00:10:02.965 "claimed": true, 00:10:02.965 "claim_type": "exclusive_write", 00:10:02.965 "zoned": false, 00:10:02.965 "supported_io_types": { 00:10:02.965 "read": true, 00:10:02.965 "write": true, 00:10:02.965 "unmap": true, 00:10:02.965 "flush": true, 00:10:02.965 "reset": true, 00:10:02.965 "nvme_admin": false, 00:10:02.965 "nvme_io": false, 00:10:02.965 "nvme_io_md": false, 00:10:02.965 "write_zeroes": true, 00:10:02.965 "zcopy": true, 00:10:02.965 "get_zone_info": false, 00:10:02.965 "zone_management": false, 00:10:02.965 "zone_append": false, 00:10:02.965 "compare": false, 00:10:02.965 "compare_and_write": false, 00:10:02.965 "abort": true, 00:10:02.965 "seek_hole": false, 00:10:02.965 "seek_data": false, 00:10:02.965 "copy": true, 00:10:02.965 "nvme_iov_md": false 00:10:02.965 }, 00:10:02.965 "memory_domains": [ 00:10:02.965 { 00:10:02.965 "dma_device_id": "system", 00:10:02.965 "dma_device_type": 1 00:10:02.965 }, 00:10:02.965 { 00:10:02.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.965 "dma_device_type": 2 00:10:02.965 } 00:10:02.965 ], 00:10:02.965 "driver_specific": {} 00:10:02.965 } 00:10:02.965 ] 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.965 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.225 "name": "Existed_Raid", 00:10:03.225 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:03.225 "strip_size_kb": 64, 00:10:03.225 "state": "configuring", 00:10:03.225 "raid_level": "concat", 00:10:03.225 "superblock": true, 00:10:03.225 "num_base_bdevs": 4, 00:10:03.225 "num_base_bdevs_discovered": 2, 00:10:03.225 "num_base_bdevs_operational": 4, 00:10:03.225 "base_bdevs_list": [ 00:10:03.225 { 00:10:03.225 "name": "BaseBdev1", 00:10:03.225 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:03.225 "is_configured": true, 00:10:03.225 "data_offset": 2048, 00:10:03.225 "data_size": 63488 00:10:03.225 }, 00:10:03.225 { 00:10:03.225 "name": "BaseBdev2", 00:10:03.225 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:03.225 "is_configured": true, 00:10:03.225 "data_offset": 2048, 00:10:03.225 "data_size": 63488 00:10:03.225 }, 00:10:03.225 { 00:10:03.225 "name": "BaseBdev3", 00:10:03.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.225 "is_configured": false, 00:10:03.225 "data_offset": 0, 00:10:03.225 "data_size": 0 00:10:03.225 }, 00:10:03.225 { 00:10:03.225 "name": "BaseBdev4", 00:10:03.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.225 "is_configured": false, 00:10:03.225 "data_offset": 0, 00:10:03.225 "data_size": 0 00:10:03.225 } 00:10:03.225 ] 00:10:03.225 }' 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.225 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.484 [2024-09-30 14:09:07.978678] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:03.484 BaseBdev3 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.484 14:09:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.484 [ 00:10:03.484 { 00:10:03.484 "name": "BaseBdev3", 00:10:03.484 "aliases": [ 00:10:03.484 "f914f36b-434a-416d-8cd4-0cb71d9ec486" 00:10:03.484 ], 00:10:03.484 "product_name": "Malloc disk", 00:10:03.484 "block_size": 512, 00:10:03.484 "num_blocks": 65536, 00:10:03.484 "uuid": "f914f36b-434a-416d-8cd4-0cb71d9ec486", 00:10:03.484 "assigned_rate_limits": { 00:10:03.484 "rw_ios_per_sec": 0, 00:10:03.484 "rw_mbytes_per_sec": 0, 00:10:03.484 "r_mbytes_per_sec": 0, 00:10:03.484 "w_mbytes_per_sec": 0 00:10:03.484 }, 00:10:03.484 "claimed": true, 00:10:03.484 "claim_type": "exclusive_write", 00:10:03.484 "zoned": false, 00:10:03.484 "supported_io_types": { 00:10:03.484 "read": true, 00:10:03.484 "write": true, 00:10:03.484 "unmap": true, 00:10:03.484 "flush": true, 00:10:03.484 "reset": true, 00:10:03.484 "nvme_admin": false, 00:10:03.484 "nvme_io": false, 00:10:03.484 "nvme_io_md": false, 00:10:03.484 "write_zeroes": true, 00:10:03.484 "zcopy": true, 00:10:03.484 "get_zone_info": false, 00:10:03.484 "zone_management": false, 00:10:03.484 "zone_append": false, 00:10:03.484 "compare": false, 00:10:03.484 "compare_and_write": false, 00:10:03.484 "abort": true, 00:10:03.484 "seek_hole": false, 00:10:03.484 "seek_data": false, 00:10:03.484 "copy": true, 00:10:03.484 "nvme_iov_md": false 00:10:03.484 }, 00:10:03.484 "memory_domains": [ 00:10:03.484 { 00:10:03.484 "dma_device_id": "system", 00:10:03.484 "dma_device_type": 1 00:10:03.484 }, 00:10:03.484 { 00:10:03.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.484 "dma_device_type": 2 00:10:03.484 } 00:10:03.484 ], 00:10:03.484 "driver_specific": {} 00:10:03.484 } 00:10:03.484 ] 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:03.484 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.485 "name": "Existed_Raid", 00:10:03.485 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:03.485 "strip_size_kb": 64, 00:10:03.485 "state": "configuring", 00:10:03.485 "raid_level": "concat", 00:10:03.485 "superblock": true, 00:10:03.485 "num_base_bdevs": 4, 00:10:03.485 "num_base_bdevs_discovered": 3, 00:10:03.485 "num_base_bdevs_operational": 4, 00:10:03.485 "base_bdevs_list": [ 00:10:03.485 { 00:10:03.485 "name": "BaseBdev1", 00:10:03.485 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:03.485 "is_configured": true, 00:10:03.485 "data_offset": 2048, 00:10:03.485 "data_size": 63488 00:10:03.485 }, 00:10:03.485 { 00:10:03.485 "name": "BaseBdev2", 00:10:03.485 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:03.485 "is_configured": true, 00:10:03.485 "data_offset": 2048, 00:10:03.485 "data_size": 63488 00:10:03.485 }, 00:10:03.485 { 00:10:03.485 "name": "BaseBdev3", 00:10:03.485 "uuid": "f914f36b-434a-416d-8cd4-0cb71d9ec486", 00:10:03.485 "is_configured": true, 00:10:03.485 "data_offset": 2048, 00:10:03.485 "data_size": 63488 00:10:03.485 }, 00:10:03.485 { 00:10:03.485 "name": "BaseBdev4", 00:10:03.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.485 "is_configured": false, 00:10:03.485 "data_offset": 0, 00:10:03.485 "data_size": 0 00:10:03.485 } 00:10:03.485 ] 00:10:03.485 }' 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.485 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.052 [2024-09-30 14:09:08.476903] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:04.052 [2024-09-30 14:09:08.477121] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:04.052 [2024-09-30 14:09:08.477145] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:04.052 [2024-09-30 14:09:08.477415] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:04.052 BaseBdev4 00:10:04.052 [2024-09-30 14:09:08.477558] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:04.052 [2024-09-30 14:09:08.477574] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:04.052 [2024-09-30 14:09:08.477701] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.052 [ 00:10:04.052 { 00:10:04.052 "name": "BaseBdev4", 00:10:04.052 "aliases": [ 00:10:04.052 "be1c175b-6b0d-4952-a8cb-e7ab14063d64" 00:10:04.052 ], 00:10:04.052 "product_name": "Malloc disk", 00:10:04.052 "block_size": 512, 00:10:04.052 "num_blocks": 65536, 00:10:04.052 "uuid": "be1c175b-6b0d-4952-a8cb-e7ab14063d64", 00:10:04.052 "assigned_rate_limits": { 00:10:04.052 "rw_ios_per_sec": 0, 00:10:04.052 "rw_mbytes_per_sec": 0, 00:10:04.052 "r_mbytes_per_sec": 0, 00:10:04.052 "w_mbytes_per_sec": 0 00:10:04.052 }, 00:10:04.052 "claimed": true, 00:10:04.052 "claim_type": "exclusive_write", 00:10:04.052 "zoned": false, 00:10:04.052 "supported_io_types": { 00:10:04.052 "read": true, 00:10:04.052 "write": true, 00:10:04.052 "unmap": true, 00:10:04.052 "flush": true, 00:10:04.052 "reset": true, 00:10:04.052 "nvme_admin": false, 00:10:04.052 "nvme_io": false, 00:10:04.052 "nvme_io_md": false, 00:10:04.052 "write_zeroes": true, 00:10:04.052 "zcopy": true, 00:10:04.052 "get_zone_info": false, 00:10:04.052 "zone_management": false, 00:10:04.052 "zone_append": false, 00:10:04.052 "compare": false, 00:10:04.052 "compare_and_write": false, 00:10:04.052 "abort": true, 00:10:04.052 "seek_hole": false, 00:10:04.052 "seek_data": false, 00:10:04.052 "copy": true, 00:10:04.052 "nvme_iov_md": false 00:10:04.052 }, 00:10:04.052 "memory_domains": [ 00:10:04.052 { 00:10:04.052 "dma_device_id": "system", 00:10:04.052 "dma_device_type": 1 00:10:04.052 }, 00:10:04.052 { 00:10:04.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.052 "dma_device_type": 2 00:10:04.052 } 00:10:04.052 ], 00:10:04.052 "driver_specific": {} 00:10:04.052 } 00:10:04.052 ] 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:04.052 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.053 "name": "Existed_Raid", 00:10:04.053 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:04.053 "strip_size_kb": 64, 00:10:04.053 "state": "online", 00:10:04.053 "raid_level": "concat", 00:10:04.053 "superblock": true, 00:10:04.053 "num_base_bdevs": 4, 00:10:04.053 "num_base_bdevs_discovered": 4, 00:10:04.053 "num_base_bdevs_operational": 4, 00:10:04.053 "base_bdevs_list": [ 00:10:04.053 { 00:10:04.053 "name": "BaseBdev1", 00:10:04.053 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:04.053 "is_configured": true, 00:10:04.053 "data_offset": 2048, 00:10:04.053 "data_size": 63488 00:10:04.053 }, 00:10:04.053 { 00:10:04.053 "name": "BaseBdev2", 00:10:04.053 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:04.053 "is_configured": true, 00:10:04.053 "data_offset": 2048, 00:10:04.053 "data_size": 63488 00:10:04.053 }, 00:10:04.053 { 00:10:04.053 "name": "BaseBdev3", 00:10:04.053 "uuid": "f914f36b-434a-416d-8cd4-0cb71d9ec486", 00:10:04.053 "is_configured": true, 00:10:04.053 "data_offset": 2048, 00:10:04.053 "data_size": 63488 00:10:04.053 }, 00:10:04.053 { 00:10:04.053 "name": "BaseBdev4", 00:10:04.053 "uuid": "be1c175b-6b0d-4952-a8cb-e7ab14063d64", 00:10:04.053 "is_configured": true, 00:10:04.053 "data_offset": 2048, 00:10:04.053 "data_size": 63488 00:10:04.053 } 00:10:04.053 ] 00:10:04.053 }' 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.053 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.311 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:04.311 [2024-09-30 14:09:08.956403] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.569 14:09:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.569 14:09:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:04.569 "name": "Existed_Raid", 00:10:04.569 "aliases": [ 00:10:04.569 "8a3c3653-aded-48c7-bca3-1abc5c2915ae" 00:10:04.569 ], 00:10:04.569 "product_name": "Raid Volume", 00:10:04.569 "block_size": 512, 00:10:04.569 "num_blocks": 253952, 00:10:04.569 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:04.569 "assigned_rate_limits": { 00:10:04.569 "rw_ios_per_sec": 0, 00:10:04.569 "rw_mbytes_per_sec": 0, 00:10:04.569 "r_mbytes_per_sec": 0, 00:10:04.569 "w_mbytes_per_sec": 0 00:10:04.569 }, 00:10:04.569 "claimed": false, 00:10:04.569 "zoned": false, 00:10:04.569 "supported_io_types": { 00:10:04.569 "read": true, 00:10:04.569 "write": true, 00:10:04.569 "unmap": true, 00:10:04.569 "flush": true, 00:10:04.569 "reset": true, 00:10:04.569 "nvme_admin": false, 00:10:04.569 "nvme_io": false, 00:10:04.569 "nvme_io_md": false, 00:10:04.569 "write_zeroes": true, 00:10:04.569 "zcopy": false, 00:10:04.569 "get_zone_info": false, 00:10:04.569 "zone_management": false, 00:10:04.569 "zone_append": false, 00:10:04.569 "compare": false, 00:10:04.569 "compare_and_write": false, 00:10:04.569 "abort": false, 00:10:04.569 "seek_hole": false, 00:10:04.569 "seek_data": false, 00:10:04.569 "copy": false, 00:10:04.569 "nvme_iov_md": false 00:10:04.569 }, 00:10:04.569 "memory_domains": [ 00:10:04.569 { 00:10:04.569 "dma_device_id": "system", 00:10:04.569 "dma_device_type": 1 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.569 "dma_device_type": 2 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "system", 00:10:04.569 "dma_device_type": 1 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.569 "dma_device_type": 2 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "system", 00:10:04.569 "dma_device_type": 1 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.569 "dma_device_type": 2 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "system", 00:10:04.569 "dma_device_type": 1 00:10:04.569 }, 00:10:04.569 { 00:10:04.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.569 "dma_device_type": 2 00:10:04.569 } 00:10:04.569 ], 00:10:04.569 "driver_specific": { 00:10:04.569 "raid": { 00:10:04.569 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:04.569 "strip_size_kb": 64, 00:10:04.569 "state": "online", 00:10:04.569 "raid_level": "concat", 00:10:04.569 "superblock": true, 00:10:04.569 "num_base_bdevs": 4, 00:10:04.569 "num_base_bdevs_discovered": 4, 00:10:04.569 "num_base_bdevs_operational": 4, 00:10:04.569 "base_bdevs_list": [ 00:10:04.569 { 00:10:04.569 "name": "BaseBdev1", 00:10:04.570 "uuid": "a7413b80-74d1-4d06-ba0b-9aa77572d575", 00:10:04.570 "is_configured": true, 00:10:04.570 "data_offset": 2048, 00:10:04.570 "data_size": 63488 00:10:04.570 }, 00:10:04.570 { 00:10:04.570 "name": "BaseBdev2", 00:10:04.570 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:04.570 "is_configured": true, 00:10:04.570 "data_offset": 2048, 00:10:04.570 "data_size": 63488 00:10:04.570 }, 00:10:04.570 { 00:10:04.570 "name": "BaseBdev3", 00:10:04.570 "uuid": "f914f36b-434a-416d-8cd4-0cb71d9ec486", 00:10:04.570 "is_configured": true, 00:10:04.570 "data_offset": 2048, 00:10:04.570 "data_size": 63488 00:10:04.570 }, 00:10:04.570 { 00:10:04.570 "name": "BaseBdev4", 00:10:04.570 "uuid": "be1c175b-6b0d-4952-a8cb-e7ab14063d64", 00:10:04.570 "is_configured": true, 00:10:04.570 "data_offset": 2048, 00:10:04.570 "data_size": 63488 00:10:04.570 } 00:10:04.570 ] 00:10:04.570 } 00:10:04.570 } 00:10:04.570 }' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:04.570 BaseBdev2 00:10:04.570 BaseBdev3 00:10:04.570 BaseBdev4' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.570 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.829 [2024-09-30 14:09:09.299592] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:04.829 [2024-09-30 14:09:09.299616] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:04.829 [2024-09-30 14:09:09.299664] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.829 "name": "Existed_Raid", 00:10:04.829 "uuid": "8a3c3653-aded-48c7-bca3-1abc5c2915ae", 00:10:04.829 "strip_size_kb": 64, 00:10:04.829 "state": "offline", 00:10:04.829 "raid_level": "concat", 00:10:04.829 "superblock": true, 00:10:04.829 "num_base_bdevs": 4, 00:10:04.829 "num_base_bdevs_discovered": 3, 00:10:04.829 "num_base_bdevs_operational": 3, 00:10:04.829 "base_bdevs_list": [ 00:10:04.829 { 00:10:04.829 "name": null, 00:10:04.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.829 "is_configured": false, 00:10:04.829 "data_offset": 0, 00:10:04.829 "data_size": 63488 00:10:04.829 }, 00:10:04.829 { 00:10:04.829 "name": "BaseBdev2", 00:10:04.829 "uuid": "4e1e3775-68a4-4275-b197-4a233258ceb5", 00:10:04.829 "is_configured": true, 00:10:04.829 "data_offset": 2048, 00:10:04.829 "data_size": 63488 00:10:04.829 }, 00:10:04.829 { 00:10:04.829 "name": "BaseBdev3", 00:10:04.829 "uuid": "f914f36b-434a-416d-8cd4-0cb71d9ec486", 00:10:04.829 "is_configured": true, 00:10:04.829 "data_offset": 2048, 00:10:04.829 "data_size": 63488 00:10:04.829 }, 00:10:04.829 { 00:10:04.829 "name": "BaseBdev4", 00:10:04.829 "uuid": "be1c175b-6b0d-4952-a8cb-e7ab14063d64", 00:10:04.829 "is_configured": true, 00:10:04.829 "data_offset": 2048, 00:10:04.829 "data_size": 63488 00:10:04.829 } 00:10:04.829 ] 00:10:04.829 }' 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.829 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.088 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:05.088 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:05.088 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:05.088 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.088 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 [2024-09-30 14:09:09.773973] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 [2024-09-30 14:09:09.844800] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.347 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.347 [2024-09-30 14:09:09.915635] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:05.347 [2024-09-30 14:09:09.915750] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.348 BaseBdev2 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.348 14:09:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.607 [ 00:10:05.607 { 00:10:05.607 "name": "BaseBdev2", 00:10:05.607 "aliases": [ 00:10:05.607 "5d07638b-892b-41ac-94d3-5fa48777a26d" 00:10:05.607 ], 00:10:05.607 "product_name": "Malloc disk", 00:10:05.607 "block_size": 512, 00:10:05.607 "num_blocks": 65536, 00:10:05.607 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:05.607 "assigned_rate_limits": { 00:10:05.607 "rw_ios_per_sec": 0, 00:10:05.607 "rw_mbytes_per_sec": 0, 00:10:05.607 "r_mbytes_per_sec": 0, 00:10:05.607 "w_mbytes_per_sec": 0 00:10:05.607 }, 00:10:05.607 "claimed": false, 00:10:05.607 "zoned": false, 00:10:05.607 "supported_io_types": { 00:10:05.607 "read": true, 00:10:05.607 "write": true, 00:10:05.607 "unmap": true, 00:10:05.607 "flush": true, 00:10:05.607 "reset": true, 00:10:05.607 "nvme_admin": false, 00:10:05.607 "nvme_io": false, 00:10:05.607 "nvme_io_md": false, 00:10:05.607 "write_zeroes": true, 00:10:05.607 "zcopy": true, 00:10:05.607 "get_zone_info": false, 00:10:05.607 "zone_management": false, 00:10:05.607 "zone_append": false, 00:10:05.607 "compare": false, 00:10:05.607 "compare_and_write": false, 00:10:05.607 "abort": true, 00:10:05.607 "seek_hole": false, 00:10:05.607 "seek_data": false, 00:10:05.607 "copy": true, 00:10:05.607 "nvme_iov_md": false 00:10:05.607 }, 00:10:05.607 "memory_domains": [ 00:10:05.607 { 00:10:05.607 "dma_device_id": "system", 00:10:05.607 "dma_device_type": 1 00:10:05.607 }, 00:10:05.607 { 00:10:05.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.607 "dma_device_type": 2 00:10:05.607 } 00:10:05.607 ], 00:10:05.607 "driver_specific": {} 00:10:05.607 } 00:10:05.607 ] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.607 BaseBdev3 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.607 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.607 [ 00:10:05.607 { 00:10:05.607 "name": "BaseBdev3", 00:10:05.607 "aliases": [ 00:10:05.607 "10513871-689b-4fed-b39d-eeff79c7e672" 00:10:05.607 ], 00:10:05.607 "product_name": "Malloc disk", 00:10:05.607 "block_size": 512, 00:10:05.607 "num_blocks": 65536, 00:10:05.607 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:05.607 "assigned_rate_limits": { 00:10:05.607 "rw_ios_per_sec": 0, 00:10:05.607 "rw_mbytes_per_sec": 0, 00:10:05.607 "r_mbytes_per_sec": 0, 00:10:05.607 "w_mbytes_per_sec": 0 00:10:05.607 }, 00:10:05.607 "claimed": false, 00:10:05.607 "zoned": false, 00:10:05.607 "supported_io_types": { 00:10:05.607 "read": true, 00:10:05.607 "write": true, 00:10:05.607 "unmap": true, 00:10:05.607 "flush": true, 00:10:05.607 "reset": true, 00:10:05.607 "nvme_admin": false, 00:10:05.607 "nvme_io": false, 00:10:05.607 "nvme_io_md": false, 00:10:05.607 "write_zeroes": true, 00:10:05.607 "zcopy": true, 00:10:05.607 "get_zone_info": false, 00:10:05.607 "zone_management": false, 00:10:05.608 "zone_append": false, 00:10:05.608 "compare": false, 00:10:05.608 "compare_and_write": false, 00:10:05.608 "abort": true, 00:10:05.608 "seek_hole": false, 00:10:05.608 "seek_data": false, 00:10:05.608 "copy": true, 00:10:05.608 "nvme_iov_md": false 00:10:05.608 }, 00:10:05.608 "memory_domains": [ 00:10:05.608 { 00:10:05.608 "dma_device_id": "system", 00:10:05.608 "dma_device_type": 1 00:10:05.608 }, 00:10:05.608 { 00:10:05.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.608 "dma_device_type": 2 00:10:05.608 } 00:10:05.608 ], 00:10:05.608 "driver_specific": {} 00:10:05.608 } 00:10:05.608 ] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.608 BaseBdev4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.608 [ 00:10:05.608 { 00:10:05.608 "name": "BaseBdev4", 00:10:05.608 "aliases": [ 00:10:05.608 "bcf9e8a1-77dc-4266-b66e-f0e990967ffc" 00:10:05.608 ], 00:10:05.608 "product_name": "Malloc disk", 00:10:05.608 "block_size": 512, 00:10:05.608 "num_blocks": 65536, 00:10:05.608 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:05.608 "assigned_rate_limits": { 00:10:05.608 "rw_ios_per_sec": 0, 00:10:05.608 "rw_mbytes_per_sec": 0, 00:10:05.608 "r_mbytes_per_sec": 0, 00:10:05.608 "w_mbytes_per_sec": 0 00:10:05.608 }, 00:10:05.608 "claimed": false, 00:10:05.608 "zoned": false, 00:10:05.608 "supported_io_types": { 00:10:05.608 "read": true, 00:10:05.608 "write": true, 00:10:05.608 "unmap": true, 00:10:05.608 "flush": true, 00:10:05.608 "reset": true, 00:10:05.608 "nvme_admin": false, 00:10:05.608 "nvme_io": false, 00:10:05.608 "nvme_io_md": false, 00:10:05.608 "write_zeroes": true, 00:10:05.608 "zcopy": true, 00:10:05.608 "get_zone_info": false, 00:10:05.608 "zone_management": false, 00:10:05.608 "zone_append": false, 00:10:05.608 "compare": false, 00:10:05.608 "compare_and_write": false, 00:10:05.608 "abort": true, 00:10:05.608 "seek_hole": false, 00:10:05.608 "seek_data": false, 00:10:05.608 "copy": true, 00:10:05.608 "nvme_iov_md": false 00:10:05.608 }, 00:10:05.608 "memory_domains": [ 00:10:05.608 { 00:10:05.608 "dma_device_id": "system", 00:10:05.608 "dma_device_type": 1 00:10:05.608 }, 00:10:05.608 { 00:10:05.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.608 "dma_device_type": 2 00:10:05.608 } 00:10:05.608 ], 00:10:05.608 "driver_specific": {} 00:10:05.608 } 00:10:05.608 ] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.608 [2024-09-30 14:09:10.146197] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:05.608 [2024-09-30 14:09:10.146245] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:05.608 [2024-09-30 14:09:10.146264] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:05.608 [2024-09-30 14:09:10.148091] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:05.608 [2024-09-30 14:09:10.148135] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.608 "name": "Existed_Raid", 00:10:05.608 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:05.608 "strip_size_kb": 64, 00:10:05.608 "state": "configuring", 00:10:05.608 "raid_level": "concat", 00:10:05.608 "superblock": true, 00:10:05.608 "num_base_bdevs": 4, 00:10:05.608 "num_base_bdevs_discovered": 3, 00:10:05.608 "num_base_bdevs_operational": 4, 00:10:05.608 "base_bdevs_list": [ 00:10:05.608 { 00:10:05.608 "name": "BaseBdev1", 00:10:05.608 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:05.608 "is_configured": false, 00:10:05.608 "data_offset": 0, 00:10:05.608 "data_size": 0 00:10:05.608 }, 00:10:05.608 { 00:10:05.608 "name": "BaseBdev2", 00:10:05.608 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:05.608 "is_configured": true, 00:10:05.608 "data_offset": 2048, 00:10:05.608 "data_size": 63488 00:10:05.608 }, 00:10:05.608 { 00:10:05.608 "name": "BaseBdev3", 00:10:05.608 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:05.608 "is_configured": true, 00:10:05.608 "data_offset": 2048, 00:10:05.608 "data_size": 63488 00:10:05.608 }, 00:10:05.608 { 00:10:05.608 "name": "BaseBdev4", 00:10:05.608 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:05.608 "is_configured": true, 00:10:05.608 "data_offset": 2048, 00:10:05.608 "data_size": 63488 00:10:05.608 } 00:10:05.608 ] 00:10:05.608 }' 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.608 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.176 [2024-09-30 14:09:10.549496] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.176 "name": "Existed_Raid", 00:10:06.176 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:06.176 "strip_size_kb": 64, 00:10:06.176 "state": "configuring", 00:10:06.176 "raid_level": "concat", 00:10:06.176 "superblock": true, 00:10:06.176 "num_base_bdevs": 4, 00:10:06.176 "num_base_bdevs_discovered": 2, 00:10:06.176 "num_base_bdevs_operational": 4, 00:10:06.176 "base_bdevs_list": [ 00:10:06.176 { 00:10:06.176 "name": "BaseBdev1", 00:10:06.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.176 "is_configured": false, 00:10:06.176 "data_offset": 0, 00:10:06.176 "data_size": 0 00:10:06.176 }, 00:10:06.176 { 00:10:06.176 "name": null, 00:10:06.176 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:06.176 "is_configured": false, 00:10:06.176 "data_offset": 0, 00:10:06.176 "data_size": 63488 00:10:06.176 }, 00:10:06.176 { 00:10:06.176 "name": "BaseBdev3", 00:10:06.176 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:06.176 "is_configured": true, 00:10:06.176 "data_offset": 2048, 00:10:06.176 "data_size": 63488 00:10:06.176 }, 00:10:06.176 { 00:10:06.176 "name": "BaseBdev4", 00:10:06.176 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:06.176 "is_configured": true, 00:10:06.176 "data_offset": 2048, 00:10:06.176 "data_size": 63488 00:10:06.176 } 00:10:06.176 ] 00:10:06.176 }' 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.176 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.435 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.435 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.435 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.435 14:09:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:06.435 14:09:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.435 [2024-09-30 14:09:11.043364] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:06.435 BaseBdev1 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.435 [ 00:10:06.435 { 00:10:06.435 "name": "BaseBdev1", 00:10:06.435 "aliases": [ 00:10:06.435 "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432" 00:10:06.435 ], 00:10:06.435 "product_name": "Malloc disk", 00:10:06.435 "block_size": 512, 00:10:06.435 "num_blocks": 65536, 00:10:06.435 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:06.435 "assigned_rate_limits": { 00:10:06.435 "rw_ios_per_sec": 0, 00:10:06.435 "rw_mbytes_per_sec": 0, 00:10:06.435 "r_mbytes_per_sec": 0, 00:10:06.435 "w_mbytes_per_sec": 0 00:10:06.435 }, 00:10:06.435 "claimed": true, 00:10:06.435 "claim_type": "exclusive_write", 00:10:06.435 "zoned": false, 00:10:06.435 "supported_io_types": { 00:10:06.435 "read": true, 00:10:06.435 "write": true, 00:10:06.435 "unmap": true, 00:10:06.435 "flush": true, 00:10:06.435 "reset": true, 00:10:06.435 "nvme_admin": false, 00:10:06.435 "nvme_io": false, 00:10:06.435 "nvme_io_md": false, 00:10:06.435 "write_zeroes": true, 00:10:06.435 "zcopy": true, 00:10:06.435 "get_zone_info": false, 00:10:06.435 "zone_management": false, 00:10:06.435 "zone_append": false, 00:10:06.435 "compare": false, 00:10:06.435 "compare_and_write": false, 00:10:06.435 "abort": true, 00:10:06.435 "seek_hole": false, 00:10:06.435 "seek_data": false, 00:10:06.435 "copy": true, 00:10:06.435 "nvme_iov_md": false 00:10:06.435 }, 00:10:06.435 "memory_domains": [ 00:10:06.435 { 00:10:06.435 "dma_device_id": "system", 00:10:06.435 "dma_device_type": 1 00:10:06.435 }, 00:10:06.435 { 00:10:06.435 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.435 "dma_device_type": 2 00:10:06.435 } 00:10:06.435 ], 00:10:06.435 "driver_specific": {} 00:10:06.435 } 00:10:06.435 ] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.435 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.695 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.695 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.695 "name": "Existed_Raid", 00:10:06.695 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:06.695 "strip_size_kb": 64, 00:10:06.695 "state": "configuring", 00:10:06.695 "raid_level": "concat", 00:10:06.695 "superblock": true, 00:10:06.695 "num_base_bdevs": 4, 00:10:06.695 "num_base_bdevs_discovered": 3, 00:10:06.695 "num_base_bdevs_operational": 4, 00:10:06.695 "base_bdevs_list": [ 00:10:06.695 { 00:10:06.695 "name": "BaseBdev1", 00:10:06.695 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:06.695 "is_configured": true, 00:10:06.695 "data_offset": 2048, 00:10:06.695 "data_size": 63488 00:10:06.695 }, 00:10:06.695 { 00:10:06.695 "name": null, 00:10:06.695 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:06.695 "is_configured": false, 00:10:06.695 "data_offset": 0, 00:10:06.695 "data_size": 63488 00:10:06.695 }, 00:10:06.695 { 00:10:06.695 "name": "BaseBdev3", 00:10:06.695 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:06.695 "is_configured": true, 00:10:06.695 "data_offset": 2048, 00:10:06.695 "data_size": 63488 00:10:06.695 }, 00:10:06.695 { 00:10:06.695 "name": "BaseBdev4", 00:10:06.695 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:06.695 "is_configured": true, 00:10:06.695 "data_offset": 2048, 00:10:06.695 "data_size": 63488 00:10:06.695 } 00:10:06.695 ] 00:10:06.695 }' 00:10:06.695 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.695 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.955 [2024-09-30 14:09:11.530623] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.955 "name": "Existed_Raid", 00:10:06.955 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:06.955 "strip_size_kb": 64, 00:10:06.955 "state": "configuring", 00:10:06.955 "raid_level": "concat", 00:10:06.955 "superblock": true, 00:10:06.955 "num_base_bdevs": 4, 00:10:06.955 "num_base_bdevs_discovered": 2, 00:10:06.955 "num_base_bdevs_operational": 4, 00:10:06.955 "base_bdevs_list": [ 00:10:06.955 { 00:10:06.955 "name": "BaseBdev1", 00:10:06.955 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:06.955 "is_configured": true, 00:10:06.955 "data_offset": 2048, 00:10:06.955 "data_size": 63488 00:10:06.955 }, 00:10:06.955 { 00:10:06.955 "name": null, 00:10:06.955 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:06.955 "is_configured": false, 00:10:06.955 "data_offset": 0, 00:10:06.955 "data_size": 63488 00:10:06.955 }, 00:10:06.955 { 00:10:06.955 "name": null, 00:10:06.955 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:06.955 "is_configured": false, 00:10:06.955 "data_offset": 0, 00:10:06.955 "data_size": 63488 00:10:06.955 }, 00:10:06.955 { 00:10:06.955 "name": "BaseBdev4", 00:10:06.955 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:06.955 "is_configured": true, 00:10:06.955 "data_offset": 2048, 00:10:06.955 "data_size": 63488 00:10:06.955 } 00:10:06.955 ] 00:10:06.955 }' 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.955 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.523 14:09:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.523 [2024-09-30 14:09:12.001825] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.523 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.524 "name": "Existed_Raid", 00:10:07.524 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:07.524 "strip_size_kb": 64, 00:10:07.524 "state": "configuring", 00:10:07.524 "raid_level": "concat", 00:10:07.524 "superblock": true, 00:10:07.524 "num_base_bdevs": 4, 00:10:07.524 "num_base_bdevs_discovered": 3, 00:10:07.524 "num_base_bdevs_operational": 4, 00:10:07.524 "base_bdevs_list": [ 00:10:07.524 { 00:10:07.524 "name": "BaseBdev1", 00:10:07.524 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:07.524 "is_configured": true, 00:10:07.524 "data_offset": 2048, 00:10:07.524 "data_size": 63488 00:10:07.524 }, 00:10:07.524 { 00:10:07.524 "name": null, 00:10:07.524 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:07.524 "is_configured": false, 00:10:07.524 "data_offset": 0, 00:10:07.524 "data_size": 63488 00:10:07.524 }, 00:10:07.524 { 00:10:07.524 "name": "BaseBdev3", 00:10:07.524 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:07.524 "is_configured": true, 00:10:07.524 "data_offset": 2048, 00:10:07.524 "data_size": 63488 00:10:07.524 }, 00:10:07.524 { 00:10:07.524 "name": "BaseBdev4", 00:10:07.524 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:07.524 "is_configured": true, 00:10:07.524 "data_offset": 2048, 00:10:07.524 "data_size": 63488 00:10:07.524 } 00:10:07.524 ] 00:10:07.524 }' 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.524 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.093 [2024-09-30 14:09:12.548915] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.093 "name": "Existed_Raid", 00:10:08.093 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:08.093 "strip_size_kb": 64, 00:10:08.093 "state": "configuring", 00:10:08.093 "raid_level": "concat", 00:10:08.093 "superblock": true, 00:10:08.093 "num_base_bdevs": 4, 00:10:08.093 "num_base_bdevs_discovered": 2, 00:10:08.093 "num_base_bdevs_operational": 4, 00:10:08.093 "base_bdevs_list": [ 00:10:08.093 { 00:10:08.093 "name": null, 00:10:08.093 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:08.093 "is_configured": false, 00:10:08.093 "data_offset": 0, 00:10:08.093 "data_size": 63488 00:10:08.093 }, 00:10:08.093 { 00:10:08.093 "name": null, 00:10:08.093 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:08.093 "is_configured": false, 00:10:08.093 "data_offset": 0, 00:10:08.093 "data_size": 63488 00:10:08.093 }, 00:10:08.093 { 00:10:08.093 "name": "BaseBdev3", 00:10:08.093 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:08.093 "is_configured": true, 00:10:08.093 "data_offset": 2048, 00:10:08.093 "data_size": 63488 00:10:08.093 }, 00:10:08.093 { 00:10:08.093 "name": "BaseBdev4", 00:10:08.093 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:08.093 "is_configured": true, 00:10:08.093 "data_offset": 2048, 00:10:08.093 "data_size": 63488 00:10:08.093 } 00:10:08.093 ] 00:10:08.093 }' 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.093 14:09:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.681 [2024-09-30 14:09:13.066473] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.681 "name": "Existed_Raid", 00:10:08.681 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:08.681 "strip_size_kb": 64, 00:10:08.681 "state": "configuring", 00:10:08.681 "raid_level": "concat", 00:10:08.681 "superblock": true, 00:10:08.681 "num_base_bdevs": 4, 00:10:08.681 "num_base_bdevs_discovered": 3, 00:10:08.681 "num_base_bdevs_operational": 4, 00:10:08.681 "base_bdevs_list": [ 00:10:08.681 { 00:10:08.681 "name": null, 00:10:08.681 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:08.681 "is_configured": false, 00:10:08.681 "data_offset": 0, 00:10:08.681 "data_size": 63488 00:10:08.681 }, 00:10:08.681 { 00:10:08.681 "name": "BaseBdev2", 00:10:08.681 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:08.681 "is_configured": true, 00:10:08.681 "data_offset": 2048, 00:10:08.681 "data_size": 63488 00:10:08.681 }, 00:10:08.681 { 00:10:08.681 "name": "BaseBdev3", 00:10:08.681 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:08.681 "is_configured": true, 00:10:08.681 "data_offset": 2048, 00:10:08.681 "data_size": 63488 00:10:08.681 }, 00:10:08.681 { 00:10:08.681 "name": "BaseBdev4", 00:10:08.681 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:08.681 "is_configured": true, 00:10:08.681 "data_offset": 2048, 00:10:08.681 "data_size": 63488 00:10:08.681 } 00:10:08.681 ] 00:10:08.681 }' 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.681 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.956 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c8fb3a23-27e2-4e4d-9b17-f8004d2aa432 00:10:08.957 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.957 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.216 [2024-09-30 14:09:13.616279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:09.216 [2024-09-30 14:09:13.616455] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:09.216 [2024-09-30 14:09:13.616470] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:09.216 NewBaseBdev 00:10:09.216 [2024-09-30 14:09:13.616740] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:10:09.216 [2024-09-30 14:09:13.616851] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:09.216 [2024-09-30 14:09:13.616860] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:09.216 [2024-09-30 14:09:13.616951] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.216 [ 00:10:09.216 { 00:10:09.216 "name": "NewBaseBdev", 00:10:09.216 "aliases": [ 00:10:09.216 "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432" 00:10:09.216 ], 00:10:09.216 "product_name": "Malloc disk", 00:10:09.216 "block_size": 512, 00:10:09.216 "num_blocks": 65536, 00:10:09.216 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:09.216 "assigned_rate_limits": { 00:10:09.216 "rw_ios_per_sec": 0, 00:10:09.216 "rw_mbytes_per_sec": 0, 00:10:09.216 "r_mbytes_per_sec": 0, 00:10:09.216 "w_mbytes_per_sec": 0 00:10:09.216 }, 00:10:09.216 "claimed": true, 00:10:09.216 "claim_type": "exclusive_write", 00:10:09.216 "zoned": false, 00:10:09.216 "supported_io_types": { 00:10:09.216 "read": true, 00:10:09.216 "write": true, 00:10:09.216 "unmap": true, 00:10:09.216 "flush": true, 00:10:09.216 "reset": true, 00:10:09.216 "nvme_admin": false, 00:10:09.216 "nvme_io": false, 00:10:09.216 "nvme_io_md": false, 00:10:09.216 "write_zeroes": true, 00:10:09.216 "zcopy": true, 00:10:09.216 "get_zone_info": false, 00:10:09.216 "zone_management": false, 00:10:09.216 "zone_append": false, 00:10:09.216 "compare": false, 00:10:09.216 "compare_and_write": false, 00:10:09.216 "abort": true, 00:10:09.216 "seek_hole": false, 00:10:09.216 "seek_data": false, 00:10:09.216 "copy": true, 00:10:09.216 "nvme_iov_md": false 00:10:09.216 }, 00:10:09.216 "memory_domains": [ 00:10:09.216 { 00:10:09.216 "dma_device_id": "system", 00:10:09.216 "dma_device_type": 1 00:10:09.216 }, 00:10:09.216 { 00:10:09.216 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.216 "dma_device_type": 2 00:10:09.216 } 00:10:09.216 ], 00:10:09.216 "driver_specific": {} 00:10:09.216 } 00:10:09.216 ] 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:09.216 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.217 "name": "Existed_Raid", 00:10:09.217 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:09.217 "strip_size_kb": 64, 00:10:09.217 "state": "online", 00:10:09.217 "raid_level": "concat", 00:10:09.217 "superblock": true, 00:10:09.217 "num_base_bdevs": 4, 00:10:09.217 "num_base_bdevs_discovered": 4, 00:10:09.217 "num_base_bdevs_operational": 4, 00:10:09.217 "base_bdevs_list": [ 00:10:09.217 { 00:10:09.217 "name": "NewBaseBdev", 00:10:09.217 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:09.217 "is_configured": true, 00:10:09.217 "data_offset": 2048, 00:10:09.217 "data_size": 63488 00:10:09.217 }, 00:10:09.217 { 00:10:09.217 "name": "BaseBdev2", 00:10:09.217 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:09.217 "is_configured": true, 00:10:09.217 "data_offset": 2048, 00:10:09.217 "data_size": 63488 00:10:09.217 }, 00:10:09.217 { 00:10:09.217 "name": "BaseBdev3", 00:10:09.217 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:09.217 "is_configured": true, 00:10:09.217 "data_offset": 2048, 00:10:09.217 "data_size": 63488 00:10:09.217 }, 00:10:09.217 { 00:10:09.217 "name": "BaseBdev4", 00:10:09.217 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:09.217 "is_configured": true, 00:10:09.217 "data_offset": 2048, 00:10:09.217 "data_size": 63488 00:10:09.217 } 00:10:09.217 ] 00:10:09.217 }' 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.217 14:09:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.476 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:09.476 [2024-09-30 14:09:14.111790] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:09.736 "name": "Existed_Raid", 00:10:09.736 "aliases": [ 00:10:09.736 "359c30cd-c5d6-469b-9550-93501c17e67b" 00:10:09.736 ], 00:10:09.736 "product_name": "Raid Volume", 00:10:09.736 "block_size": 512, 00:10:09.736 "num_blocks": 253952, 00:10:09.736 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:09.736 "assigned_rate_limits": { 00:10:09.736 "rw_ios_per_sec": 0, 00:10:09.736 "rw_mbytes_per_sec": 0, 00:10:09.736 "r_mbytes_per_sec": 0, 00:10:09.736 "w_mbytes_per_sec": 0 00:10:09.736 }, 00:10:09.736 "claimed": false, 00:10:09.736 "zoned": false, 00:10:09.736 "supported_io_types": { 00:10:09.736 "read": true, 00:10:09.736 "write": true, 00:10:09.736 "unmap": true, 00:10:09.736 "flush": true, 00:10:09.736 "reset": true, 00:10:09.736 "nvme_admin": false, 00:10:09.736 "nvme_io": false, 00:10:09.736 "nvme_io_md": false, 00:10:09.736 "write_zeroes": true, 00:10:09.736 "zcopy": false, 00:10:09.736 "get_zone_info": false, 00:10:09.736 "zone_management": false, 00:10:09.736 "zone_append": false, 00:10:09.736 "compare": false, 00:10:09.736 "compare_and_write": false, 00:10:09.736 "abort": false, 00:10:09.736 "seek_hole": false, 00:10:09.736 "seek_data": false, 00:10:09.736 "copy": false, 00:10:09.736 "nvme_iov_md": false 00:10:09.736 }, 00:10:09.736 "memory_domains": [ 00:10:09.736 { 00:10:09.736 "dma_device_id": "system", 00:10:09.736 "dma_device_type": 1 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.736 "dma_device_type": 2 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "system", 00:10:09.736 "dma_device_type": 1 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.736 "dma_device_type": 2 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "system", 00:10:09.736 "dma_device_type": 1 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.736 "dma_device_type": 2 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "system", 00:10:09.736 "dma_device_type": 1 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.736 "dma_device_type": 2 00:10:09.736 } 00:10:09.736 ], 00:10:09.736 "driver_specific": { 00:10:09.736 "raid": { 00:10:09.736 "uuid": "359c30cd-c5d6-469b-9550-93501c17e67b", 00:10:09.736 "strip_size_kb": 64, 00:10:09.736 "state": "online", 00:10:09.736 "raid_level": "concat", 00:10:09.736 "superblock": true, 00:10:09.736 "num_base_bdevs": 4, 00:10:09.736 "num_base_bdevs_discovered": 4, 00:10:09.736 "num_base_bdevs_operational": 4, 00:10:09.736 "base_bdevs_list": [ 00:10:09.736 { 00:10:09.736 "name": "NewBaseBdev", 00:10:09.736 "uuid": "c8fb3a23-27e2-4e4d-9b17-f8004d2aa432", 00:10:09.736 "is_configured": true, 00:10:09.736 "data_offset": 2048, 00:10:09.736 "data_size": 63488 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "name": "BaseBdev2", 00:10:09.736 "uuid": "5d07638b-892b-41ac-94d3-5fa48777a26d", 00:10:09.736 "is_configured": true, 00:10:09.736 "data_offset": 2048, 00:10:09.736 "data_size": 63488 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "name": "BaseBdev3", 00:10:09.736 "uuid": "10513871-689b-4fed-b39d-eeff79c7e672", 00:10:09.736 "is_configured": true, 00:10:09.736 "data_offset": 2048, 00:10:09.736 "data_size": 63488 00:10:09.736 }, 00:10:09.736 { 00:10:09.736 "name": "BaseBdev4", 00:10:09.736 "uuid": "bcf9e8a1-77dc-4266-b66e-f0e990967ffc", 00:10:09.736 "is_configured": true, 00:10:09.736 "data_offset": 2048, 00:10:09.736 "data_size": 63488 00:10:09.736 } 00:10:09.736 ] 00:10:09.736 } 00:10:09.736 } 00:10:09.736 }' 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:09.736 BaseBdev2 00:10:09.736 BaseBdev3 00:10:09.736 BaseBdev4' 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.736 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.737 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.996 [2024-09-30 14:09:14.454856] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:09.996 [2024-09-30 14:09:14.454920] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:09.996 [2024-09-30 14:09:14.455021] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:09.996 [2024-09-30 14:09:14.455100] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:09.996 [2024-09-30 14:09:14.455162] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 83845 00:10:09.996 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 83845 ']' 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 83845 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83845 00:10:09.997 killing process with pid 83845 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83845' 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 83845 00:10:09.997 [2024-09-30 14:09:14.502243] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:09.997 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 83845 00:10:09.997 [2024-09-30 14:09:14.541664] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:10.256 14:09:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:10.256 00:10:10.256 real 0m9.636s 00:10:10.256 user 0m16.435s 00:10:10.256 sys 0m2.020s 00:10:10.256 ************************************ 00:10:10.256 END TEST raid_state_function_test_sb 00:10:10.256 ************************************ 00:10:10.256 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:10.256 14:09:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.256 14:09:14 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:10.256 14:09:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:10.256 14:09:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:10.256 14:09:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:10.256 ************************************ 00:10:10.256 START TEST raid_superblock_test 00:10:10.256 ************************************ 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 4 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=84493 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 84493 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 84493 ']' 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:10.256 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:10.256 14:09:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.516 [2024-09-30 14:09:14.951121] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:10.516 [2024-09-30 14:09:14.951294] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84493 ] 00:10:10.516 [2024-09-30 14:09:15.082238] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:10.516 [2024-09-30 14:09:15.112215] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:10.516 [2024-09-30 14:09:15.157345] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:10.776 [2024-09-30 14:09:15.198613] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:10.776 [2024-09-30 14:09:15.198651] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.347 malloc1 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.347 [2024-09-30 14:09:15.792321] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:11.347 [2024-09-30 14:09:15.792443] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.347 [2024-09-30 14:09:15.792488] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:11.347 [2024-09-30 14:09:15.792544] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.347 [2024-09-30 14:09:15.794684] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.347 [2024-09-30 14:09:15.794750] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:11.347 pt1 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.347 malloc2 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.347 [2024-09-30 14:09:15.838113] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:11.347 [2024-09-30 14:09:15.838269] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.347 [2024-09-30 14:09:15.838306] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:11.347 [2024-09-30 14:09:15.838326] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.347 [2024-09-30 14:09:15.842813] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.347 [2024-09-30 14:09:15.842879] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:11.347 pt2 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.347 malloc3 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:11.347 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.348 [2024-09-30 14:09:15.868587] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:11.348 [2024-09-30 14:09:15.868672] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.348 [2024-09-30 14:09:15.868704] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:11.348 [2024-09-30 14:09:15.868730] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.348 [2024-09-30 14:09:15.870726] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.348 [2024-09-30 14:09:15.870795] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:11.348 pt3 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.348 malloc4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.348 [2024-09-30 14:09:15.900861] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:11.348 [2024-09-30 14:09:15.900943] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.348 [2024-09-30 14:09:15.900979] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:11.348 [2024-09-30 14:09:15.901006] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.348 [2024-09-30 14:09:15.903005] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.348 [2024-09-30 14:09:15.903069] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:11.348 pt4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.348 [2024-09-30 14:09:15.912887] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:11.348 [2024-09-30 14:09:15.914652] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:11.348 [2024-09-30 14:09:15.914715] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:11.348 [2024-09-30 14:09:15.914795] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:11.348 [2024-09-30 14:09:15.914939] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:10:11.348 [2024-09-30 14:09:15.914950] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:11.348 [2024-09-30 14:09:15.915192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:11.348 [2024-09-30 14:09:15.915312] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:10:11.348 [2024-09-30 14:09:15.915324] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:10:11.348 [2024-09-30 14:09:15.915428] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.348 "name": "raid_bdev1", 00:10:11.348 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:11.348 "strip_size_kb": 64, 00:10:11.348 "state": "online", 00:10:11.348 "raid_level": "concat", 00:10:11.348 "superblock": true, 00:10:11.348 "num_base_bdevs": 4, 00:10:11.348 "num_base_bdevs_discovered": 4, 00:10:11.348 "num_base_bdevs_operational": 4, 00:10:11.348 "base_bdevs_list": [ 00:10:11.348 { 00:10:11.348 "name": "pt1", 00:10:11.348 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:11.348 "is_configured": true, 00:10:11.348 "data_offset": 2048, 00:10:11.348 "data_size": 63488 00:10:11.348 }, 00:10:11.348 { 00:10:11.348 "name": "pt2", 00:10:11.348 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.348 "is_configured": true, 00:10:11.348 "data_offset": 2048, 00:10:11.348 "data_size": 63488 00:10:11.348 }, 00:10:11.348 { 00:10:11.348 "name": "pt3", 00:10:11.348 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.348 "is_configured": true, 00:10:11.348 "data_offset": 2048, 00:10:11.348 "data_size": 63488 00:10:11.348 }, 00:10:11.348 { 00:10:11.348 "name": "pt4", 00:10:11.348 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.348 "is_configured": true, 00:10:11.348 "data_offset": 2048, 00:10:11.348 "data_size": 63488 00:10:11.348 } 00:10:11.348 ] 00:10:11.348 }' 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.348 14:09:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.917 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:11.917 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:11.917 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:11.917 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:11.917 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.918 [2024-09-30 14:09:16.344438] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:11.918 "name": "raid_bdev1", 00:10:11.918 "aliases": [ 00:10:11.918 "a3b1d666-137d-4b85-bc48-08586f723850" 00:10:11.918 ], 00:10:11.918 "product_name": "Raid Volume", 00:10:11.918 "block_size": 512, 00:10:11.918 "num_blocks": 253952, 00:10:11.918 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:11.918 "assigned_rate_limits": { 00:10:11.918 "rw_ios_per_sec": 0, 00:10:11.918 "rw_mbytes_per_sec": 0, 00:10:11.918 "r_mbytes_per_sec": 0, 00:10:11.918 "w_mbytes_per_sec": 0 00:10:11.918 }, 00:10:11.918 "claimed": false, 00:10:11.918 "zoned": false, 00:10:11.918 "supported_io_types": { 00:10:11.918 "read": true, 00:10:11.918 "write": true, 00:10:11.918 "unmap": true, 00:10:11.918 "flush": true, 00:10:11.918 "reset": true, 00:10:11.918 "nvme_admin": false, 00:10:11.918 "nvme_io": false, 00:10:11.918 "nvme_io_md": false, 00:10:11.918 "write_zeroes": true, 00:10:11.918 "zcopy": false, 00:10:11.918 "get_zone_info": false, 00:10:11.918 "zone_management": false, 00:10:11.918 "zone_append": false, 00:10:11.918 "compare": false, 00:10:11.918 "compare_and_write": false, 00:10:11.918 "abort": false, 00:10:11.918 "seek_hole": false, 00:10:11.918 "seek_data": false, 00:10:11.918 "copy": false, 00:10:11.918 "nvme_iov_md": false 00:10:11.918 }, 00:10:11.918 "memory_domains": [ 00:10:11.918 { 00:10:11.918 "dma_device_id": "system", 00:10:11.918 "dma_device_type": 1 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.918 "dma_device_type": 2 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "system", 00:10:11.918 "dma_device_type": 1 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.918 "dma_device_type": 2 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "system", 00:10:11.918 "dma_device_type": 1 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.918 "dma_device_type": 2 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "system", 00:10:11.918 "dma_device_type": 1 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.918 "dma_device_type": 2 00:10:11.918 } 00:10:11.918 ], 00:10:11.918 "driver_specific": { 00:10:11.918 "raid": { 00:10:11.918 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:11.918 "strip_size_kb": 64, 00:10:11.918 "state": "online", 00:10:11.918 "raid_level": "concat", 00:10:11.918 "superblock": true, 00:10:11.918 "num_base_bdevs": 4, 00:10:11.918 "num_base_bdevs_discovered": 4, 00:10:11.918 "num_base_bdevs_operational": 4, 00:10:11.918 "base_bdevs_list": [ 00:10:11.918 { 00:10:11.918 "name": "pt1", 00:10:11.918 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:11.918 "is_configured": true, 00:10:11.918 "data_offset": 2048, 00:10:11.918 "data_size": 63488 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "name": "pt2", 00:10:11.918 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.918 "is_configured": true, 00:10:11.918 "data_offset": 2048, 00:10:11.918 "data_size": 63488 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "name": "pt3", 00:10:11.918 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.918 "is_configured": true, 00:10:11.918 "data_offset": 2048, 00:10:11.918 "data_size": 63488 00:10:11.918 }, 00:10:11.918 { 00:10:11.918 "name": "pt4", 00:10:11.918 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.918 "is_configured": true, 00:10:11.918 "data_offset": 2048, 00:10:11.918 "data_size": 63488 00:10:11.918 } 00:10:11.918 ] 00:10:11.918 } 00:10:11.918 } 00:10:11.918 }' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:11.918 pt2 00:10:11.918 pt3 00:10:11.918 pt4' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.918 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 [2024-09-30 14:09:16.651831] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a3b1d666-137d-4b85-bc48-08586f723850 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a3b1d666-137d-4b85-bc48-08586f723850 ']' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 [2024-09-30 14:09:16.699507] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:12.178 [2024-09-30 14:09:16.699529] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:12.178 [2024-09-30 14:09:16.699604] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.178 [2024-09-30 14:09:16.699673] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:12.178 [2024-09-30 14:09:16.699685] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.178 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.438 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.438 [2024-09-30 14:09:16.863234] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:12.438 [2024-09-30 14:09:16.865092] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:12.438 [2024-09-30 14:09:16.865171] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:12.438 [2024-09-30 14:09:16.865218] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:12.438 [2024-09-30 14:09:16.865279] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:12.438 [2024-09-30 14:09:16.865346] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:12.438 [2024-09-30 14:09:16.865408] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:12.438 [2024-09-30 14:09:16.865448] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:12.438 [2024-09-30 14:09:16.865504] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:12.438 [2024-09-30 14:09:16.865542] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:10:12.438 request: 00:10:12.438 { 00:10:12.438 "name": "raid_bdev1", 00:10:12.438 "raid_level": "concat", 00:10:12.438 "base_bdevs": [ 00:10:12.438 "malloc1", 00:10:12.438 "malloc2", 00:10:12.438 "malloc3", 00:10:12.438 "malloc4" 00:10:12.438 ], 00:10:12.438 "strip_size_kb": 64, 00:10:12.438 "superblock": false, 00:10:12.438 "method": "bdev_raid_create", 00:10:12.438 "req_id": 1 00:10:12.438 } 00:10:12.438 Got JSON-RPC error response 00:10:12.439 response: 00:10:12.439 { 00:10:12.439 "code": -17, 00:10:12.439 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:12.439 } 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.439 [2024-09-30 14:09:16.919110] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:12.439 [2024-09-30 14:09:16.919191] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:12.439 [2024-09-30 14:09:16.919237] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:12.439 [2024-09-30 14:09:16.919266] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:12.439 [2024-09-30 14:09:16.921300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:12.439 [2024-09-30 14:09:16.921367] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:12.439 [2024-09-30 14:09:16.921447] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:12.439 [2024-09-30 14:09:16.921520] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:12.439 pt1 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.439 "name": "raid_bdev1", 00:10:12.439 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:12.439 "strip_size_kb": 64, 00:10:12.439 "state": "configuring", 00:10:12.439 "raid_level": "concat", 00:10:12.439 "superblock": true, 00:10:12.439 "num_base_bdevs": 4, 00:10:12.439 "num_base_bdevs_discovered": 1, 00:10:12.439 "num_base_bdevs_operational": 4, 00:10:12.439 "base_bdevs_list": [ 00:10:12.439 { 00:10:12.439 "name": "pt1", 00:10:12.439 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:12.439 "is_configured": true, 00:10:12.439 "data_offset": 2048, 00:10:12.439 "data_size": 63488 00:10:12.439 }, 00:10:12.439 { 00:10:12.439 "name": null, 00:10:12.439 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:12.439 "is_configured": false, 00:10:12.439 "data_offset": 2048, 00:10:12.439 "data_size": 63488 00:10:12.439 }, 00:10:12.439 { 00:10:12.439 "name": null, 00:10:12.439 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:12.439 "is_configured": false, 00:10:12.439 "data_offset": 2048, 00:10:12.439 "data_size": 63488 00:10:12.439 }, 00:10:12.439 { 00:10:12.439 "name": null, 00:10:12.439 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:12.439 "is_configured": false, 00:10:12.439 "data_offset": 2048, 00:10:12.439 "data_size": 63488 00:10:12.439 } 00:10:12.439 ] 00:10:12.439 }' 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.439 14:09:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.008 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.009 [2024-09-30 14:09:17.386338] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:13.009 [2024-09-30 14:09:17.386394] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.009 [2024-09-30 14:09:17.386426] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:13.009 [2024-09-30 14:09:17.386436] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.009 [2024-09-30 14:09:17.386834] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.009 [2024-09-30 14:09:17.386864] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:13.009 [2024-09-30 14:09:17.386929] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:13.009 [2024-09-30 14:09:17.386953] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:13.009 pt2 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.009 [2024-09-30 14:09:17.398332] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.009 "name": "raid_bdev1", 00:10:13.009 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:13.009 "strip_size_kb": 64, 00:10:13.009 "state": "configuring", 00:10:13.009 "raid_level": "concat", 00:10:13.009 "superblock": true, 00:10:13.009 "num_base_bdevs": 4, 00:10:13.009 "num_base_bdevs_discovered": 1, 00:10:13.009 "num_base_bdevs_operational": 4, 00:10:13.009 "base_bdevs_list": [ 00:10:13.009 { 00:10:13.009 "name": "pt1", 00:10:13.009 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:13.009 "is_configured": true, 00:10:13.009 "data_offset": 2048, 00:10:13.009 "data_size": 63488 00:10:13.009 }, 00:10:13.009 { 00:10:13.009 "name": null, 00:10:13.009 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:13.009 "is_configured": false, 00:10:13.009 "data_offset": 0, 00:10:13.009 "data_size": 63488 00:10:13.009 }, 00:10:13.009 { 00:10:13.009 "name": null, 00:10:13.009 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:13.009 "is_configured": false, 00:10:13.009 "data_offset": 2048, 00:10:13.009 "data_size": 63488 00:10:13.009 }, 00:10:13.009 { 00:10:13.009 "name": null, 00:10:13.009 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:13.009 "is_configured": false, 00:10:13.009 "data_offset": 2048, 00:10:13.009 "data_size": 63488 00:10:13.009 } 00:10:13.009 ] 00:10:13.009 }' 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.009 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.269 [2024-09-30 14:09:17.809630] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:13.269 [2024-09-30 14:09:17.809727] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.269 [2024-09-30 14:09:17.809761] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:13.269 [2024-09-30 14:09:17.809789] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.269 [2024-09-30 14:09:17.810181] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.269 [2024-09-30 14:09:17.810238] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:13.269 [2024-09-30 14:09:17.810333] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:13.269 [2024-09-30 14:09:17.810390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:13.269 pt2 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.269 [2024-09-30 14:09:17.821582] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:13.269 [2024-09-30 14:09:17.821660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.269 [2024-09-30 14:09:17.821718] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:13.269 [2024-09-30 14:09:17.821743] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.269 [2024-09-30 14:09:17.822069] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.269 [2024-09-30 14:09:17.822124] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:13.269 [2024-09-30 14:09:17.822202] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:13.269 [2024-09-30 14:09:17.822248] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:13.269 pt3 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.269 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.269 [2024-09-30 14:09:17.833572] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:13.269 [2024-09-30 14:09:17.833647] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.269 [2024-09-30 14:09:17.833680] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:13.269 [2024-09-30 14:09:17.833705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.269 [2024-09-30 14:09:17.833996] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.269 [2024-09-30 14:09:17.834051] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:13.269 [2024-09-30 14:09:17.834139] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:13.269 [2024-09-30 14:09:17.834189] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:13.269 [2024-09-30 14:09:17.834319] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:13.269 [2024-09-30 14:09:17.834355] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:13.270 [2024-09-30 14:09:17.834610] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:10:13.270 [2024-09-30 14:09:17.834751] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:13.270 [2024-09-30 14:09:17.834818] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:10:13.270 [2024-09-30 14:09:17.834942] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.270 pt4 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.270 "name": "raid_bdev1", 00:10:13.270 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:13.270 "strip_size_kb": 64, 00:10:13.270 "state": "online", 00:10:13.270 "raid_level": "concat", 00:10:13.270 "superblock": true, 00:10:13.270 "num_base_bdevs": 4, 00:10:13.270 "num_base_bdevs_discovered": 4, 00:10:13.270 "num_base_bdevs_operational": 4, 00:10:13.270 "base_bdevs_list": [ 00:10:13.270 { 00:10:13.270 "name": "pt1", 00:10:13.270 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:13.270 "is_configured": true, 00:10:13.270 "data_offset": 2048, 00:10:13.270 "data_size": 63488 00:10:13.270 }, 00:10:13.270 { 00:10:13.270 "name": "pt2", 00:10:13.270 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:13.270 "is_configured": true, 00:10:13.270 "data_offset": 2048, 00:10:13.270 "data_size": 63488 00:10:13.270 }, 00:10:13.270 { 00:10:13.270 "name": "pt3", 00:10:13.270 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:13.270 "is_configured": true, 00:10:13.270 "data_offset": 2048, 00:10:13.270 "data_size": 63488 00:10:13.270 }, 00:10:13.270 { 00:10:13.270 "name": "pt4", 00:10:13.270 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:13.270 "is_configured": true, 00:10:13.270 "data_offset": 2048, 00:10:13.270 "data_size": 63488 00:10:13.270 } 00:10:13.270 ] 00:10:13.270 }' 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.270 14:09:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.837 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:13.837 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:13.837 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:13.837 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:13.837 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:13.838 [2024-09-30 14:09:18.301102] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:13.838 "name": "raid_bdev1", 00:10:13.838 "aliases": [ 00:10:13.838 "a3b1d666-137d-4b85-bc48-08586f723850" 00:10:13.838 ], 00:10:13.838 "product_name": "Raid Volume", 00:10:13.838 "block_size": 512, 00:10:13.838 "num_blocks": 253952, 00:10:13.838 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:13.838 "assigned_rate_limits": { 00:10:13.838 "rw_ios_per_sec": 0, 00:10:13.838 "rw_mbytes_per_sec": 0, 00:10:13.838 "r_mbytes_per_sec": 0, 00:10:13.838 "w_mbytes_per_sec": 0 00:10:13.838 }, 00:10:13.838 "claimed": false, 00:10:13.838 "zoned": false, 00:10:13.838 "supported_io_types": { 00:10:13.838 "read": true, 00:10:13.838 "write": true, 00:10:13.838 "unmap": true, 00:10:13.838 "flush": true, 00:10:13.838 "reset": true, 00:10:13.838 "nvme_admin": false, 00:10:13.838 "nvme_io": false, 00:10:13.838 "nvme_io_md": false, 00:10:13.838 "write_zeroes": true, 00:10:13.838 "zcopy": false, 00:10:13.838 "get_zone_info": false, 00:10:13.838 "zone_management": false, 00:10:13.838 "zone_append": false, 00:10:13.838 "compare": false, 00:10:13.838 "compare_and_write": false, 00:10:13.838 "abort": false, 00:10:13.838 "seek_hole": false, 00:10:13.838 "seek_data": false, 00:10:13.838 "copy": false, 00:10:13.838 "nvme_iov_md": false 00:10:13.838 }, 00:10:13.838 "memory_domains": [ 00:10:13.838 { 00:10:13.838 "dma_device_id": "system", 00:10:13.838 "dma_device_type": 1 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.838 "dma_device_type": 2 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "system", 00:10:13.838 "dma_device_type": 1 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.838 "dma_device_type": 2 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "system", 00:10:13.838 "dma_device_type": 1 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.838 "dma_device_type": 2 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "system", 00:10:13.838 "dma_device_type": 1 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.838 "dma_device_type": 2 00:10:13.838 } 00:10:13.838 ], 00:10:13.838 "driver_specific": { 00:10:13.838 "raid": { 00:10:13.838 "uuid": "a3b1d666-137d-4b85-bc48-08586f723850", 00:10:13.838 "strip_size_kb": 64, 00:10:13.838 "state": "online", 00:10:13.838 "raid_level": "concat", 00:10:13.838 "superblock": true, 00:10:13.838 "num_base_bdevs": 4, 00:10:13.838 "num_base_bdevs_discovered": 4, 00:10:13.838 "num_base_bdevs_operational": 4, 00:10:13.838 "base_bdevs_list": [ 00:10:13.838 { 00:10:13.838 "name": "pt1", 00:10:13.838 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:13.838 "is_configured": true, 00:10:13.838 "data_offset": 2048, 00:10:13.838 "data_size": 63488 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "name": "pt2", 00:10:13.838 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:13.838 "is_configured": true, 00:10:13.838 "data_offset": 2048, 00:10:13.838 "data_size": 63488 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "name": "pt3", 00:10:13.838 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:13.838 "is_configured": true, 00:10:13.838 "data_offset": 2048, 00:10:13.838 "data_size": 63488 00:10:13.838 }, 00:10:13.838 { 00:10:13.838 "name": "pt4", 00:10:13.838 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:13.838 "is_configured": true, 00:10:13.838 "data_offset": 2048, 00:10:13.838 "data_size": 63488 00:10:13.838 } 00:10:13.838 ] 00:10:13.838 } 00:10:13.838 } 00:10:13.838 }' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:13.838 pt2 00:10:13.838 pt3 00:10:13.838 pt4' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.838 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:14.098 [2024-09-30 14:09:18.612510] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a3b1d666-137d-4b85-bc48-08586f723850 '!=' a3b1d666-137d-4b85-bc48-08586f723850 ']' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 84493 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 84493 ']' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 84493 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84493 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84493' 00:10:14.098 killing process with pid 84493 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 84493 00:10:14.098 [2024-09-30 14:09:18.694676] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:14.098 [2024-09-30 14:09:18.694811] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:14.098 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 84493 00:10:14.098 [2024-09-30 14:09:18.694920] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:14.098 [2024-09-30 14:09:18.694939] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:10:14.098 [2024-09-30 14:09:18.736707] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:14.358 14:09:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:14.358 00:10:14.358 real 0m4.118s 00:10:14.358 user 0m6.433s 00:10:14.358 sys 0m0.961s 00:10:14.358 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:14.358 14:09:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.358 ************************************ 00:10:14.358 END TEST raid_superblock_test 00:10:14.358 ************************************ 00:10:14.618 14:09:19 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:14.618 14:09:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:14.618 14:09:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:14.618 14:09:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:14.618 ************************************ 00:10:14.618 START TEST raid_read_error_test 00:10:14.618 ************************************ 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 read 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.SrhaiI55if 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=84747 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 84747 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 84747 ']' 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:14.618 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:14.618 14:09:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.618 [2024-09-30 14:09:19.173349] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:14.618 [2024-09-30 14:09:19.173473] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84747 ] 00:10:14.877 [2024-09-30 14:09:19.305991] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:14.877 [2024-09-30 14:09:19.332497] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:14.877 [2024-09-30 14:09:19.377968] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:14.877 [2024-09-30 14:09:19.419162] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:14.877 [2024-09-30 14:09:19.419198] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 BaseBdev1_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 true 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 [2024-09-30 14:09:20.048453] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:15.446 [2024-09-30 14:09:20.048519] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.446 [2024-09-30 14:09:20.048539] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:15.446 [2024-09-30 14:09:20.048552] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.446 [2024-09-30 14:09:20.050610] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.446 [2024-09-30 14:09:20.050689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:15.446 BaseBdev1 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 BaseBdev2_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 true 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.446 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.446 [2024-09-30 14:09:20.099549] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:15.446 [2024-09-30 14:09:20.099640] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.446 [2024-09-30 14:09:20.099673] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:15.446 [2024-09-30 14:09:20.099703] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.446 [2024-09-30 14:09:20.101759] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.446 [2024-09-30 14:09:20.101835] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:15.706 BaseBdev2 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 BaseBdev3_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 true 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 [2024-09-30 14:09:20.139979] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:15.706 [2024-09-30 14:09:20.140030] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.706 [2024-09-30 14:09:20.140048] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:15.706 [2024-09-30 14:09:20.140058] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.706 [2024-09-30 14:09:20.142049] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.706 [2024-09-30 14:09:20.142087] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:15.706 BaseBdev3 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 BaseBdev4_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 true 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 [2024-09-30 14:09:20.180320] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:15.706 [2024-09-30 14:09:20.180373] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.706 [2024-09-30 14:09:20.180390] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:15.706 [2024-09-30 14:09:20.180416] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.706 [2024-09-30 14:09:20.182444] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.706 [2024-09-30 14:09:20.182494] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:15.706 BaseBdev4 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.706 [2024-09-30 14:09:20.192362] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:15.706 [2024-09-30 14:09:20.194164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:15.706 [2024-09-30 14:09:20.194273] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:15.706 [2024-09-30 14:09:20.194330] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:15.706 [2024-09-30 14:09:20.194516] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:10:15.706 [2024-09-30 14:09:20.194537] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:15.706 [2024-09-30 14:09:20.194747] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:15.706 [2024-09-30 14:09:20.194893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:10:15.706 [2024-09-30 14:09:20.194903] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:10:15.706 [2024-09-30 14:09:20.195032] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:15.706 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.707 "name": "raid_bdev1", 00:10:15.707 "uuid": "8dc4ef2d-bcba-4c2e-8979-8cd4655ec804", 00:10:15.707 "strip_size_kb": 64, 00:10:15.707 "state": "online", 00:10:15.707 "raid_level": "concat", 00:10:15.707 "superblock": true, 00:10:15.707 "num_base_bdevs": 4, 00:10:15.707 "num_base_bdevs_discovered": 4, 00:10:15.707 "num_base_bdevs_operational": 4, 00:10:15.707 "base_bdevs_list": [ 00:10:15.707 { 00:10:15.707 "name": "BaseBdev1", 00:10:15.707 "uuid": "f9df8577-e222-55ef-b733-4f1461f28066", 00:10:15.707 "is_configured": true, 00:10:15.707 "data_offset": 2048, 00:10:15.707 "data_size": 63488 00:10:15.707 }, 00:10:15.707 { 00:10:15.707 "name": "BaseBdev2", 00:10:15.707 "uuid": "aef916ba-2a17-58e9-84a4-15c085eaa993", 00:10:15.707 "is_configured": true, 00:10:15.707 "data_offset": 2048, 00:10:15.707 "data_size": 63488 00:10:15.707 }, 00:10:15.707 { 00:10:15.707 "name": "BaseBdev3", 00:10:15.707 "uuid": "ec197cd0-69b4-54b6-802a-51df2484eb07", 00:10:15.707 "is_configured": true, 00:10:15.707 "data_offset": 2048, 00:10:15.707 "data_size": 63488 00:10:15.707 }, 00:10:15.707 { 00:10:15.707 "name": "BaseBdev4", 00:10:15.707 "uuid": "6c8f75c5-6285-5c5a-b625-7cc1983cf31d", 00:10:15.707 "is_configured": true, 00:10:15.707 "data_offset": 2048, 00:10:15.707 "data_size": 63488 00:10:15.707 } 00:10:15.707 ] 00:10:15.707 }' 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.707 14:09:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:16.276 14:09:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:16.276 [2024-09-30 14:09:20.735809] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.217 "name": "raid_bdev1", 00:10:17.217 "uuid": "8dc4ef2d-bcba-4c2e-8979-8cd4655ec804", 00:10:17.217 "strip_size_kb": 64, 00:10:17.217 "state": "online", 00:10:17.217 "raid_level": "concat", 00:10:17.217 "superblock": true, 00:10:17.217 "num_base_bdevs": 4, 00:10:17.217 "num_base_bdevs_discovered": 4, 00:10:17.217 "num_base_bdevs_operational": 4, 00:10:17.217 "base_bdevs_list": [ 00:10:17.217 { 00:10:17.217 "name": "BaseBdev1", 00:10:17.217 "uuid": "f9df8577-e222-55ef-b733-4f1461f28066", 00:10:17.217 "is_configured": true, 00:10:17.217 "data_offset": 2048, 00:10:17.217 "data_size": 63488 00:10:17.217 }, 00:10:17.217 { 00:10:17.217 "name": "BaseBdev2", 00:10:17.217 "uuid": "aef916ba-2a17-58e9-84a4-15c085eaa993", 00:10:17.217 "is_configured": true, 00:10:17.217 "data_offset": 2048, 00:10:17.217 "data_size": 63488 00:10:17.217 }, 00:10:17.217 { 00:10:17.217 "name": "BaseBdev3", 00:10:17.217 "uuid": "ec197cd0-69b4-54b6-802a-51df2484eb07", 00:10:17.217 "is_configured": true, 00:10:17.217 "data_offset": 2048, 00:10:17.217 "data_size": 63488 00:10:17.217 }, 00:10:17.217 { 00:10:17.217 "name": "BaseBdev4", 00:10:17.217 "uuid": "6c8f75c5-6285-5c5a-b625-7cc1983cf31d", 00:10:17.217 "is_configured": true, 00:10:17.217 "data_offset": 2048, 00:10:17.217 "data_size": 63488 00:10:17.217 } 00:10:17.217 ] 00:10:17.217 }' 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.217 14:09:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.477 [2024-09-30 14:09:22.115229] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:17.477 [2024-09-30 14:09:22.115265] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:17.477 [2024-09-30 14:09:22.117686] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:17.477 [2024-09-30 14:09:22.117741] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.477 [2024-09-30 14:09:22.117785] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:17.477 [2024-09-30 14:09:22.117796] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:10:17.477 { 00:10:17.477 "results": [ 00:10:17.477 { 00:10:17.477 "job": "raid_bdev1", 00:10:17.477 "core_mask": "0x1", 00:10:17.477 "workload": "randrw", 00:10:17.477 "percentage": 50, 00:10:17.477 "status": "finished", 00:10:17.477 "queue_depth": 1, 00:10:17.477 "io_size": 131072, 00:10:17.477 "runtime": 1.380319, 00:10:17.477 "iops": 17259.77835558302, 00:10:17.477 "mibps": 2157.4722944478776, 00:10:17.477 "io_failed": 1, 00:10:17.477 "io_timeout": 0, 00:10:17.477 "avg_latency_us": 80.33230896685714, 00:10:17.477 "min_latency_us": 24.705676855895195, 00:10:17.477 "max_latency_us": 1359.3711790393013 00:10:17.477 } 00:10:17.477 ], 00:10:17.477 "core_count": 1 00:10:17.477 } 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 84747 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 84747 ']' 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 84747 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:17.477 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84747 00:10:17.738 killing process with pid 84747 00:10:17.738 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:17.738 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:17.738 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84747' 00:10:17.738 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 84747 00:10:17.738 [2024-09-30 14:09:22.160807] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:17.738 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 84747 00:10:17.738 [2024-09-30 14:09:22.195228] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.SrhaiI55if 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:17.998 00:10:17.998 real 0m3.384s 00:10:17.998 user 0m4.251s 00:10:17.998 sys 0m0.585s 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:17.998 ************************************ 00:10:17.998 END TEST raid_read_error_test 00:10:17.998 ************************************ 00:10:17.998 14:09:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.998 14:09:22 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:17.998 14:09:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:17.998 14:09:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:17.998 14:09:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:17.998 ************************************ 00:10:17.998 START TEST raid_write_error_test 00:10:17.998 ************************************ 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 write 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:17.998 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.bgxewdfKDg 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=84876 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 84876 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 84876 ']' 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:17.999 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:17.999 14:09:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.999 [2024-09-30 14:09:22.630047] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:17.999 [2024-09-30 14:09:22.630169] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84876 ] 00:10:18.259 [2024-09-30 14:09:22.767393] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:18.259 [2024-09-30 14:09:22.795937] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:18.259 [2024-09-30 14:09:22.840114] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:18.259 [2024-09-30 14:09:22.881173] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:18.259 [2024-09-30 14:09:22.881217] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.826 BaseBdev1_malloc 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.826 true 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.826 [2024-09-30 14:09:23.474459] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:18.826 [2024-09-30 14:09:23.474531] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:18.826 [2024-09-30 14:09:23.474554] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:18.826 [2024-09-30 14:09:23.474573] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:18.826 [2024-09-30 14:09:23.476652] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:18.826 [2024-09-30 14:09:23.476691] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:18.826 BaseBdev1 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.826 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 BaseBdev2_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 true 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 [2024-09-30 14:09:23.525280] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:19.086 [2024-09-30 14:09:23.525329] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.086 [2024-09-30 14:09:23.525345] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:19.086 [2024-09-30 14:09:23.525355] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.086 [2024-09-30 14:09:23.527409] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.086 [2024-09-30 14:09:23.527450] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:19.086 BaseBdev2 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 BaseBdev3_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 true 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 [2024-09-30 14:09:23.565490] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:19.086 [2024-09-30 14:09:23.565533] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.086 [2024-09-30 14:09:23.565548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:19.086 [2024-09-30 14:09:23.565557] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.086 [2024-09-30 14:09:23.567553] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.086 [2024-09-30 14:09:23.567640] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:19.086 BaseBdev3 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 BaseBdev4_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 true 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 [2024-09-30 14:09:23.605788] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:19.086 [2024-09-30 14:09:23.605838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.086 [2024-09-30 14:09:23.605854] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:19.086 [2024-09-30 14:09:23.605864] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.086 [2024-09-30 14:09:23.607891] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.086 [2024-09-30 14:09:23.607930] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:19.086 BaseBdev4 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 [2024-09-30 14:09:23.617839] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:19.086 [2024-09-30 14:09:23.619682] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:19.086 [2024-09-30 14:09:23.619747] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:19.086 [2024-09-30 14:09:23.619800] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:19.086 [2024-09-30 14:09:23.619981] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:10:19.086 [2024-09-30 14:09:23.619997] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:19.086 [2024-09-30 14:09:23.620215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:19.086 [2024-09-30 14:09:23.620335] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:10:19.086 [2024-09-30 14:09:23.620345] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:10:19.086 [2024-09-30 14:09:23.620474] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.086 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.086 "name": "raid_bdev1", 00:10:19.086 "uuid": "b4cdd8bf-8292-4f74-8be7-44de768dfbc5", 00:10:19.086 "strip_size_kb": 64, 00:10:19.086 "state": "online", 00:10:19.086 "raid_level": "concat", 00:10:19.086 "superblock": true, 00:10:19.086 "num_base_bdevs": 4, 00:10:19.086 "num_base_bdevs_discovered": 4, 00:10:19.086 "num_base_bdevs_operational": 4, 00:10:19.086 "base_bdevs_list": [ 00:10:19.086 { 00:10:19.086 "name": "BaseBdev1", 00:10:19.086 "uuid": "d61673ab-cdcf-5756-a73e-3f9c8826a1d2", 00:10:19.086 "is_configured": true, 00:10:19.086 "data_offset": 2048, 00:10:19.086 "data_size": 63488 00:10:19.086 }, 00:10:19.086 { 00:10:19.086 "name": "BaseBdev2", 00:10:19.086 "uuid": "29aac68a-f65a-5308-ab78-8210c99e2f9e", 00:10:19.086 "is_configured": true, 00:10:19.086 "data_offset": 2048, 00:10:19.086 "data_size": 63488 00:10:19.086 }, 00:10:19.086 { 00:10:19.086 "name": "BaseBdev3", 00:10:19.086 "uuid": "73d69e80-bf32-51aa-84bd-e1e6c73129b5", 00:10:19.086 "is_configured": true, 00:10:19.086 "data_offset": 2048, 00:10:19.086 "data_size": 63488 00:10:19.086 }, 00:10:19.086 { 00:10:19.087 "name": "BaseBdev4", 00:10:19.087 "uuid": "4db5cbfe-77b2-5b67-a022-8e497e8a6c61", 00:10:19.087 "is_configured": true, 00:10:19.087 "data_offset": 2048, 00:10:19.087 "data_size": 63488 00:10:19.087 } 00:10:19.087 ] 00:10:19.087 }' 00:10:19.087 14:09:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.087 14:09:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.655 14:09:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:19.655 14:09:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:19.655 [2024-09-30 14:09:24.185220] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:10:20.593 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.594 "name": "raid_bdev1", 00:10:20.594 "uuid": "b4cdd8bf-8292-4f74-8be7-44de768dfbc5", 00:10:20.594 "strip_size_kb": 64, 00:10:20.594 "state": "online", 00:10:20.594 "raid_level": "concat", 00:10:20.594 "superblock": true, 00:10:20.594 "num_base_bdevs": 4, 00:10:20.594 "num_base_bdevs_discovered": 4, 00:10:20.594 "num_base_bdevs_operational": 4, 00:10:20.594 "base_bdevs_list": [ 00:10:20.594 { 00:10:20.594 "name": "BaseBdev1", 00:10:20.594 "uuid": "d61673ab-cdcf-5756-a73e-3f9c8826a1d2", 00:10:20.594 "is_configured": true, 00:10:20.594 "data_offset": 2048, 00:10:20.594 "data_size": 63488 00:10:20.594 }, 00:10:20.594 { 00:10:20.594 "name": "BaseBdev2", 00:10:20.594 "uuid": "29aac68a-f65a-5308-ab78-8210c99e2f9e", 00:10:20.594 "is_configured": true, 00:10:20.594 "data_offset": 2048, 00:10:20.594 "data_size": 63488 00:10:20.594 }, 00:10:20.594 { 00:10:20.594 "name": "BaseBdev3", 00:10:20.594 "uuid": "73d69e80-bf32-51aa-84bd-e1e6c73129b5", 00:10:20.594 "is_configured": true, 00:10:20.594 "data_offset": 2048, 00:10:20.594 "data_size": 63488 00:10:20.594 }, 00:10:20.594 { 00:10:20.594 "name": "BaseBdev4", 00:10:20.594 "uuid": "4db5cbfe-77b2-5b67-a022-8e497e8a6c61", 00:10:20.594 "is_configured": true, 00:10:20.594 "data_offset": 2048, 00:10:20.594 "data_size": 63488 00:10:20.594 } 00:10:20.594 ] 00:10:20.594 }' 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.594 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.162 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:21.162 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.162 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.162 [2024-09-30 14:09:25.548858] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:21.162 [2024-09-30 14:09:25.548898] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:21.162 [2024-09-30 14:09:25.551438] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:21.162 [2024-09-30 14:09:25.551510] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:21.162 [2024-09-30 14:09:25.551553] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:21.162 [2024-09-30 14:09:25.551564] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:10:21.162 { 00:10:21.162 "results": [ 00:10:21.162 { 00:10:21.162 "job": "raid_bdev1", 00:10:21.162 "core_mask": "0x1", 00:10:21.162 "workload": "randrw", 00:10:21.162 "percentage": 50, 00:10:21.162 "status": "finished", 00:10:21.163 "queue_depth": 1, 00:10:21.163 "io_size": 131072, 00:10:21.163 "runtime": 1.364431, 00:10:21.163 "iops": 17097.96977641229, 00:10:21.163 "mibps": 2137.2462220515363, 00:10:21.163 "io_failed": 1, 00:10:21.163 "io_timeout": 0, 00:10:21.163 "avg_latency_us": 81.18445946426533, 00:10:21.163 "min_latency_us": 24.258515283842794, 00:10:21.163 "max_latency_us": 1352.216593886463 00:10:21.163 } 00:10:21.163 ], 00:10:21.163 "core_count": 1 00:10:21.163 } 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 84876 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 84876 ']' 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 84876 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84876 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:21.163 killing process with pid 84876 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84876' 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 84876 00:10:21.163 [2024-09-30 14:09:25.594117] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:21.163 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 84876 00:10:21.163 [2024-09-30 14:09:25.628166] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.bgxewdfKDg 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:21.422 ************************************ 00:10:21.422 END TEST raid_write_error_test 00:10:21.422 ************************************ 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:21.422 00:10:21.422 real 0m3.353s 00:10:21.422 user 0m4.211s 00:10:21.422 sys 0m0.555s 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:21.422 14:09:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.422 14:09:25 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:21.422 14:09:25 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:21.422 14:09:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:21.422 14:09:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:21.422 14:09:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:21.422 ************************************ 00:10:21.422 START TEST raid_state_function_test 00:10:21.422 ************************************ 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 false 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=85008 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:21.422 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 85008' 00:10:21.422 Process raid pid: 85008 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 85008 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 85008 ']' 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:21.423 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:21.423 14:09:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.423 [2024-09-30 14:09:26.045187] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:21.423 [2024-09-30 14:09:26.045329] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:21.681 [2024-09-30 14:09:26.184296] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:21.681 [2024-09-30 14:09:26.211632] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:21.681 [2024-09-30 14:09:26.257813] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:21.681 [2024-09-30 14:09:26.299648] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:21.681 [2024-09-30 14:09:26.299682] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:22.251 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:22.251 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:10:22.251 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:22.251 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.252 [2024-09-30 14:09:26.865155] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:22.252 [2024-09-30 14:09:26.865210] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:22.252 [2024-09-30 14:09:26.865222] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:22.252 [2024-09-30 14:09:26.865229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:22.252 [2024-09-30 14:09:26.865242] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:22.252 [2024-09-30 14:09:26.865249] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:22.252 [2024-09-30 14:09:26.865256] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:22.252 [2024-09-30 14:09:26.865263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.252 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.526 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.526 "name": "Existed_Raid", 00:10:22.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.526 "strip_size_kb": 0, 00:10:22.526 "state": "configuring", 00:10:22.526 "raid_level": "raid1", 00:10:22.526 "superblock": false, 00:10:22.526 "num_base_bdevs": 4, 00:10:22.526 "num_base_bdevs_discovered": 0, 00:10:22.526 "num_base_bdevs_operational": 4, 00:10:22.526 "base_bdevs_list": [ 00:10:22.526 { 00:10:22.526 "name": "BaseBdev1", 00:10:22.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.526 "is_configured": false, 00:10:22.526 "data_offset": 0, 00:10:22.526 "data_size": 0 00:10:22.526 }, 00:10:22.526 { 00:10:22.526 "name": "BaseBdev2", 00:10:22.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.526 "is_configured": false, 00:10:22.526 "data_offset": 0, 00:10:22.526 "data_size": 0 00:10:22.526 }, 00:10:22.526 { 00:10:22.526 "name": "BaseBdev3", 00:10:22.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.526 "is_configured": false, 00:10:22.526 "data_offset": 0, 00:10:22.526 "data_size": 0 00:10:22.526 }, 00:10:22.526 { 00:10:22.526 "name": "BaseBdev4", 00:10:22.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.526 "is_configured": false, 00:10:22.526 "data_offset": 0, 00:10:22.526 "data_size": 0 00:10:22.526 } 00:10:22.526 ] 00:10:22.526 }' 00:10:22.526 14:09:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.526 14:09:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.798 [2024-09-30 14:09:27.244430] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:22.798 [2024-09-30 14:09:27.244546] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.798 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.798 [2024-09-30 14:09:27.252421] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:22.798 [2024-09-30 14:09:27.252521] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:22.798 [2024-09-30 14:09:27.252573] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:22.798 [2024-09-30 14:09:27.252596] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:22.798 [2024-09-30 14:09:27.252654] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:22.798 [2024-09-30 14:09:27.252674] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:22.798 [2024-09-30 14:09:27.252746] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:22.798 [2024-09-30 14:09:27.252774] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.799 [2024-09-30 14:09:27.269125] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:22.799 BaseBdev1 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.799 [ 00:10:22.799 { 00:10:22.799 "name": "BaseBdev1", 00:10:22.799 "aliases": [ 00:10:22.799 "a0a31fe3-6694-4f81-bea6-c50269659069" 00:10:22.799 ], 00:10:22.799 "product_name": "Malloc disk", 00:10:22.799 "block_size": 512, 00:10:22.799 "num_blocks": 65536, 00:10:22.799 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:22.799 "assigned_rate_limits": { 00:10:22.799 "rw_ios_per_sec": 0, 00:10:22.799 "rw_mbytes_per_sec": 0, 00:10:22.799 "r_mbytes_per_sec": 0, 00:10:22.799 "w_mbytes_per_sec": 0 00:10:22.799 }, 00:10:22.799 "claimed": true, 00:10:22.799 "claim_type": "exclusive_write", 00:10:22.799 "zoned": false, 00:10:22.799 "supported_io_types": { 00:10:22.799 "read": true, 00:10:22.799 "write": true, 00:10:22.799 "unmap": true, 00:10:22.799 "flush": true, 00:10:22.799 "reset": true, 00:10:22.799 "nvme_admin": false, 00:10:22.799 "nvme_io": false, 00:10:22.799 "nvme_io_md": false, 00:10:22.799 "write_zeroes": true, 00:10:22.799 "zcopy": true, 00:10:22.799 "get_zone_info": false, 00:10:22.799 "zone_management": false, 00:10:22.799 "zone_append": false, 00:10:22.799 "compare": false, 00:10:22.799 "compare_and_write": false, 00:10:22.799 "abort": true, 00:10:22.799 "seek_hole": false, 00:10:22.799 "seek_data": false, 00:10:22.799 "copy": true, 00:10:22.799 "nvme_iov_md": false 00:10:22.799 }, 00:10:22.799 "memory_domains": [ 00:10:22.799 { 00:10:22.799 "dma_device_id": "system", 00:10:22.799 "dma_device_type": 1 00:10:22.799 }, 00:10:22.799 { 00:10:22.799 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.799 "dma_device_type": 2 00:10:22.799 } 00:10:22.799 ], 00:10:22.799 "driver_specific": {} 00:10:22.799 } 00:10:22.799 ] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.799 "name": "Existed_Raid", 00:10:22.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.799 "strip_size_kb": 0, 00:10:22.799 "state": "configuring", 00:10:22.799 "raid_level": "raid1", 00:10:22.799 "superblock": false, 00:10:22.799 "num_base_bdevs": 4, 00:10:22.799 "num_base_bdevs_discovered": 1, 00:10:22.799 "num_base_bdevs_operational": 4, 00:10:22.799 "base_bdevs_list": [ 00:10:22.799 { 00:10:22.799 "name": "BaseBdev1", 00:10:22.799 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:22.799 "is_configured": true, 00:10:22.799 "data_offset": 0, 00:10:22.799 "data_size": 65536 00:10:22.799 }, 00:10:22.799 { 00:10:22.799 "name": "BaseBdev2", 00:10:22.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.799 "is_configured": false, 00:10:22.799 "data_offset": 0, 00:10:22.799 "data_size": 0 00:10:22.799 }, 00:10:22.799 { 00:10:22.799 "name": "BaseBdev3", 00:10:22.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.799 "is_configured": false, 00:10:22.799 "data_offset": 0, 00:10:22.799 "data_size": 0 00:10:22.799 }, 00:10:22.799 { 00:10:22.799 "name": "BaseBdev4", 00:10:22.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.799 "is_configured": false, 00:10:22.799 "data_offset": 0, 00:10:22.799 "data_size": 0 00:10:22.799 } 00:10:22.799 ] 00:10:22.799 }' 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.799 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.370 [2024-09-30 14:09:27.748345] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:23.370 [2024-09-30 14:09:27.748397] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.370 [2024-09-30 14:09:27.756343] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:23.370 [2024-09-30 14:09:27.758149] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:23.370 [2024-09-30 14:09:27.758184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:23.370 [2024-09-30 14:09:27.758194] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:23.370 [2024-09-30 14:09:27.758202] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:23.370 [2024-09-30 14:09:27.758209] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:23.370 [2024-09-30 14:09:27.758215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.370 "name": "Existed_Raid", 00:10:23.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.370 "strip_size_kb": 0, 00:10:23.370 "state": "configuring", 00:10:23.370 "raid_level": "raid1", 00:10:23.370 "superblock": false, 00:10:23.370 "num_base_bdevs": 4, 00:10:23.370 "num_base_bdevs_discovered": 1, 00:10:23.370 "num_base_bdevs_operational": 4, 00:10:23.370 "base_bdevs_list": [ 00:10:23.370 { 00:10:23.370 "name": "BaseBdev1", 00:10:23.370 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:23.370 "is_configured": true, 00:10:23.370 "data_offset": 0, 00:10:23.370 "data_size": 65536 00:10:23.370 }, 00:10:23.370 { 00:10:23.370 "name": "BaseBdev2", 00:10:23.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.370 "is_configured": false, 00:10:23.370 "data_offset": 0, 00:10:23.370 "data_size": 0 00:10:23.370 }, 00:10:23.370 { 00:10:23.370 "name": "BaseBdev3", 00:10:23.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.370 "is_configured": false, 00:10:23.370 "data_offset": 0, 00:10:23.370 "data_size": 0 00:10:23.370 }, 00:10:23.370 { 00:10:23.370 "name": "BaseBdev4", 00:10:23.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.370 "is_configured": false, 00:10:23.370 "data_offset": 0, 00:10:23.370 "data_size": 0 00:10:23.370 } 00:10:23.370 ] 00:10:23.370 }' 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.370 14:09:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.630 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.631 [2024-09-30 14:09:28.214976] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:23.631 BaseBdev2 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.631 [ 00:10:23.631 { 00:10:23.631 "name": "BaseBdev2", 00:10:23.631 "aliases": [ 00:10:23.631 "5051e183-8dcc-43d7-8e84-ce0a5014ecf3" 00:10:23.631 ], 00:10:23.631 "product_name": "Malloc disk", 00:10:23.631 "block_size": 512, 00:10:23.631 "num_blocks": 65536, 00:10:23.631 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:23.631 "assigned_rate_limits": { 00:10:23.631 "rw_ios_per_sec": 0, 00:10:23.631 "rw_mbytes_per_sec": 0, 00:10:23.631 "r_mbytes_per_sec": 0, 00:10:23.631 "w_mbytes_per_sec": 0 00:10:23.631 }, 00:10:23.631 "claimed": true, 00:10:23.631 "claim_type": "exclusive_write", 00:10:23.631 "zoned": false, 00:10:23.631 "supported_io_types": { 00:10:23.631 "read": true, 00:10:23.631 "write": true, 00:10:23.631 "unmap": true, 00:10:23.631 "flush": true, 00:10:23.631 "reset": true, 00:10:23.631 "nvme_admin": false, 00:10:23.631 "nvme_io": false, 00:10:23.631 "nvme_io_md": false, 00:10:23.631 "write_zeroes": true, 00:10:23.631 "zcopy": true, 00:10:23.631 "get_zone_info": false, 00:10:23.631 "zone_management": false, 00:10:23.631 "zone_append": false, 00:10:23.631 "compare": false, 00:10:23.631 "compare_and_write": false, 00:10:23.631 "abort": true, 00:10:23.631 "seek_hole": false, 00:10:23.631 "seek_data": false, 00:10:23.631 "copy": true, 00:10:23.631 "nvme_iov_md": false 00:10:23.631 }, 00:10:23.631 "memory_domains": [ 00:10:23.631 { 00:10:23.631 "dma_device_id": "system", 00:10:23.631 "dma_device_type": 1 00:10:23.631 }, 00:10:23.631 { 00:10:23.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.631 "dma_device_type": 2 00:10:23.631 } 00:10:23.631 ], 00:10:23.631 "driver_specific": {} 00:10:23.631 } 00:10:23.631 ] 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.631 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.891 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.891 "name": "Existed_Raid", 00:10:23.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.891 "strip_size_kb": 0, 00:10:23.891 "state": "configuring", 00:10:23.891 "raid_level": "raid1", 00:10:23.891 "superblock": false, 00:10:23.891 "num_base_bdevs": 4, 00:10:23.891 "num_base_bdevs_discovered": 2, 00:10:23.891 "num_base_bdevs_operational": 4, 00:10:23.891 "base_bdevs_list": [ 00:10:23.891 { 00:10:23.891 "name": "BaseBdev1", 00:10:23.891 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:23.891 "is_configured": true, 00:10:23.891 "data_offset": 0, 00:10:23.891 "data_size": 65536 00:10:23.891 }, 00:10:23.891 { 00:10:23.891 "name": "BaseBdev2", 00:10:23.891 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:23.891 "is_configured": true, 00:10:23.891 "data_offset": 0, 00:10:23.891 "data_size": 65536 00:10:23.891 }, 00:10:23.891 { 00:10:23.891 "name": "BaseBdev3", 00:10:23.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.891 "is_configured": false, 00:10:23.891 "data_offset": 0, 00:10:23.891 "data_size": 0 00:10:23.891 }, 00:10:23.891 { 00:10:23.891 "name": "BaseBdev4", 00:10:23.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.891 "is_configured": false, 00:10:23.891 "data_offset": 0, 00:10:23.891 "data_size": 0 00:10:23.891 } 00:10:23.891 ] 00:10:23.891 }' 00:10:23.891 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.891 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.152 [2024-09-30 14:09:28.720974] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:24.152 BaseBdev3 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.152 [ 00:10:24.152 { 00:10:24.152 "name": "BaseBdev3", 00:10:24.152 "aliases": [ 00:10:24.152 "42343bfa-d180-44d6-92b4-ccb0a97395e7" 00:10:24.152 ], 00:10:24.152 "product_name": "Malloc disk", 00:10:24.152 "block_size": 512, 00:10:24.152 "num_blocks": 65536, 00:10:24.152 "uuid": "42343bfa-d180-44d6-92b4-ccb0a97395e7", 00:10:24.152 "assigned_rate_limits": { 00:10:24.152 "rw_ios_per_sec": 0, 00:10:24.152 "rw_mbytes_per_sec": 0, 00:10:24.152 "r_mbytes_per_sec": 0, 00:10:24.152 "w_mbytes_per_sec": 0 00:10:24.152 }, 00:10:24.152 "claimed": true, 00:10:24.152 "claim_type": "exclusive_write", 00:10:24.152 "zoned": false, 00:10:24.152 "supported_io_types": { 00:10:24.152 "read": true, 00:10:24.152 "write": true, 00:10:24.152 "unmap": true, 00:10:24.152 "flush": true, 00:10:24.152 "reset": true, 00:10:24.152 "nvme_admin": false, 00:10:24.152 "nvme_io": false, 00:10:24.152 "nvme_io_md": false, 00:10:24.152 "write_zeroes": true, 00:10:24.152 "zcopy": true, 00:10:24.152 "get_zone_info": false, 00:10:24.152 "zone_management": false, 00:10:24.152 "zone_append": false, 00:10:24.152 "compare": false, 00:10:24.152 "compare_and_write": false, 00:10:24.152 "abort": true, 00:10:24.152 "seek_hole": false, 00:10:24.152 "seek_data": false, 00:10:24.152 "copy": true, 00:10:24.152 "nvme_iov_md": false 00:10:24.152 }, 00:10:24.152 "memory_domains": [ 00:10:24.152 { 00:10:24.152 "dma_device_id": "system", 00:10:24.152 "dma_device_type": 1 00:10:24.152 }, 00:10:24.152 { 00:10:24.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.152 "dma_device_type": 2 00:10:24.152 } 00:10:24.152 ], 00:10:24.152 "driver_specific": {} 00:10:24.152 } 00:10:24.152 ] 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.152 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.412 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.412 "name": "Existed_Raid", 00:10:24.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.412 "strip_size_kb": 0, 00:10:24.412 "state": "configuring", 00:10:24.412 "raid_level": "raid1", 00:10:24.412 "superblock": false, 00:10:24.412 "num_base_bdevs": 4, 00:10:24.412 "num_base_bdevs_discovered": 3, 00:10:24.412 "num_base_bdevs_operational": 4, 00:10:24.412 "base_bdevs_list": [ 00:10:24.412 { 00:10:24.412 "name": "BaseBdev1", 00:10:24.412 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:24.412 "is_configured": true, 00:10:24.412 "data_offset": 0, 00:10:24.412 "data_size": 65536 00:10:24.412 }, 00:10:24.412 { 00:10:24.412 "name": "BaseBdev2", 00:10:24.412 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:24.412 "is_configured": true, 00:10:24.412 "data_offset": 0, 00:10:24.412 "data_size": 65536 00:10:24.412 }, 00:10:24.412 { 00:10:24.412 "name": "BaseBdev3", 00:10:24.412 "uuid": "42343bfa-d180-44d6-92b4-ccb0a97395e7", 00:10:24.412 "is_configured": true, 00:10:24.412 "data_offset": 0, 00:10:24.412 "data_size": 65536 00:10:24.412 }, 00:10:24.412 { 00:10:24.412 "name": "BaseBdev4", 00:10:24.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.412 "is_configured": false, 00:10:24.412 "data_offset": 0, 00:10:24.412 "data_size": 0 00:10:24.412 } 00:10:24.412 ] 00:10:24.412 }' 00:10:24.412 14:09:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.412 14:09:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.672 [2024-09-30 14:09:29.191132] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:24.672 [2024-09-30 14:09:29.191179] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:24.672 [2024-09-30 14:09:29.191198] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:24.672 [2024-09-30 14:09:29.191470] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:24.672 [2024-09-30 14:09:29.191641] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:24.672 [2024-09-30 14:09:29.191658] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:24.672 [2024-09-30 14:09:29.191888] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:24.672 BaseBdev4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.672 [ 00:10:24.672 { 00:10:24.672 "name": "BaseBdev4", 00:10:24.672 "aliases": [ 00:10:24.672 "0edb7ad7-3bca-44fe-8026-1e1be8c524d7" 00:10:24.672 ], 00:10:24.672 "product_name": "Malloc disk", 00:10:24.672 "block_size": 512, 00:10:24.672 "num_blocks": 65536, 00:10:24.672 "uuid": "0edb7ad7-3bca-44fe-8026-1e1be8c524d7", 00:10:24.672 "assigned_rate_limits": { 00:10:24.672 "rw_ios_per_sec": 0, 00:10:24.672 "rw_mbytes_per_sec": 0, 00:10:24.672 "r_mbytes_per_sec": 0, 00:10:24.672 "w_mbytes_per_sec": 0 00:10:24.672 }, 00:10:24.672 "claimed": true, 00:10:24.672 "claim_type": "exclusive_write", 00:10:24.672 "zoned": false, 00:10:24.672 "supported_io_types": { 00:10:24.672 "read": true, 00:10:24.672 "write": true, 00:10:24.672 "unmap": true, 00:10:24.672 "flush": true, 00:10:24.672 "reset": true, 00:10:24.672 "nvme_admin": false, 00:10:24.672 "nvme_io": false, 00:10:24.672 "nvme_io_md": false, 00:10:24.672 "write_zeroes": true, 00:10:24.672 "zcopy": true, 00:10:24.672 "get_zone_info": false, 00:10:24.672 "zone_management": false, 00:10:24.672 "zone_append": false, 00:10:24.672 "compare": false, 00:10:24.672 "compare_and_write": false, 00:10:24.672 "abort": true, 00:10:24.672 "seek_hole": false, 00:10:24.672 "seek_data": false, 00:10:24.672 "copy": true, 00:10:24.672 "nvme_iov_md": false 00:10:24.672 }, 00:10:24.672 "memory_domains": [ 00:10:24.672 { 00:10:24.672 "dma_device_id": "system", 00:10:24.672 "dma_device_type": 1 00:10:24.672 }, 00:10:24.672 { 00:10:24.672 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.672 "dma_device_type": 2 00:10:24.672 } 00:10:24.672 ], 00:10:24.672 "driver_specific": {} 00:10:24.672 } 00:10:24.672 ] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.672 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.672 "name": "Existed_Raid", 00:10:24.672 "uuid": "7a4a2b40-ea7e-41b0-be60-8bea30f80cec", 00:10:24.672 "strip_size_kb": 0, 00:10:24.672 "state": "online", 00:10:24.672 "raid_level": "raid1", 00:10:24.672 "superblock": false, 00:10:24.672 "num_base_bdevs": 4, 00:10:24.672 "num_base_bdevs_discovered": 4, 00:10:24.672 "num_base_bdevs_operational": 4, 00:10:24.672 "base_bdevs_list": [ 00:10:24.672 { 00:10:24.672 "name": "BaseBdev1", 00:10:24.672 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:24.672 "is_configured": true, 00:10:24.672 "data_offset": 0, 00:10:24.672 "data_size": 65536 00:10:24.672 }, 00:10:24.672 { 00:10:24.672 "name": "BaseBdev2", 00:10:24.672 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:24.672 "is_configured": true, 00:10:24.672 "data_offset": 0, 00:10:24.672 "data_size": 65536 00:10:24.672 }, 00:10:24.672 { 00:10:24.672 "name": "BaseBdev3", 00:10:24.672 "uuid": "42343bfa-d180-44d6-92b4-ccb0a97395e7", 00:10:24.672 "is_configured": true, 00:10:24.672 "data_offset": 0, 00:10:24.672 "data_size": 65536 00:10:24.672 }, 00:10:24.672 { 00:10:24.672 "name": "BaseBdev4", 00:10:24.673 "uuid": "0edb7ad7-3bca-44fe-8026-1e1be8c524d7", 00:10:24.673 "is_configured": true, 00:10:24.673 "data_offset": 0, 00:10:24.673 "data_size": 65536 00:10:24.673 } 00:10:24.673 ] 00:10:24.673 }' 00:10:24.673 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.673 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.241 [2024-09-30 14:09:29.642683] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.241 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:25.241 "name": "Existed_Raid", 00:10:25.241 "aliases": [ 00:10:25.241 "7a4a2b40-ea7e-41b0-be60-8bea30f80cec" 00:10:25.241 ], 00:10:25.241 "product_name": "Raid Volume", 00:10:25.241 "block_size": 512, 00:10:25.241 "num_blocks": 65536, 00:10:25.241 "uuid": "7a4a2b40-ea7e-41b0-be60-8bea30f80cec", 00:10:25.241 "assigned_rate_limits": { 00:10:25.241 "rw_ios_per_sec": 0, 00:10:25.242 "rw_mbytes_per_sec": 0, 00:10:25.242 "r_mbytes_per_sec": 0, 00:10:25.242 "w_mbytes_per_sec": 0 00:10:25.242 }, 00:10:25.242 "claimed": false, 00:10:25.242 "zoned": false, 00:10:25.242 "supported_io_types": { 00:10:25.242 "read": true, 00:10:25.242 "write": true, 00:10:25.242 "unmap": false, 00:10:25.242 "flush": false, 00:10:25.242 "reset": true, 00:10:25.242 "nvme_admin": false, 00:10:25.242 "nvme_io": false, 00:10:25.242 "nvme_io_md": false, 00:10:25.242 "write_zeroes": true, 00:10:25.242 "zcopy": false, 00:10:25.242 "get_zone_info": false, 00:10:25.242 "zone_management": false, 00:10:25.242 "zone_append": false, 00:10:25.242 "compare": false, 00:10:25.242 "compare_and_write": false, 00:10:25.242 "abort": false, 00:10:25.242 "seek_hole": false, 00:10:25.242 "seek_data": false, 00:10:25.242 "copy": false, 00:10:25.242 "nvme_iov_md": false 00:10:25.242 }, 00:10:25.242 "memory_domains": [ 00:10:25.242 { 00:10:25.242 "dma_device_id": "system", 00:10:25.242 "dma_device_type": 1 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.242 "dma_device_type": 2 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "system", 00:10:25.242 "dma_device_type": 1 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.242 "dma_device_type": 2 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "system", 00:10:25.242 "dma_device_type": 1 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.242 "dma_device_type": 2 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "system", 00:10:25.242 "dma_device_type": 1 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.242 "dma_device_type": 2 00:10:25.242 } 00:10:25.242 ], 00:10:25.242 "driver_specific": { 00:10:25.242 "raid": { 00:10:25.242 "uuid": "7a4a2b40-ea7e-41b0-be60-8bea30f80cec", 00:10:25.242 "strip_size_kb": 0, 00:10:25.242 "state": "online", 00:10:25.242 "raid_level": "raid1", 00:10:25.242 "superblock": false, 00:10:25.242 "num_base_bdevs": 4, 00:10:25.242 "num_base_bdevs_discovered": 4, 00:10:25.242 "num_base_bdevs_operational": 4, 00:10:25.242 "base_bdevs_list": [ 00:10:25.242 { 00:10:25.242 "name": "BaseBdev1", 00:10:25.242 "uuid": "a0a31fe3-6694-4f81-bea6-c50269659069", 00:10:25.242 "is_configured": true, 00:10:25.242 "data_offset": 0, 00:10:25.242 "data_size": 65536 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "name": "BaseBdev2", 00:10:25.242 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:25.242 "is_configured": true, 00:10:25.242 "data_offset": 0, 00:10:25.242 "data_size": 65536 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "name": "BaseBdev3", 00:10:25.242 "uuid": "42343bfa-d180-44d6-92b4-ccb0a97395e7", 00:10:25.242 "is_configured": true, 00:10:25.242 "data_offset": 0, 00:10:25.242 "data_size": 65536 00:10:25.242 }, 00:10:25.242 { 00:10:25.242 "name": "BaseBdev4", 00:10:25.242 "uuid": "0edb7ad7-3bca-44fe-8026-1e1be8c524d7", 00:10:25.242 "is_configured": true, 00:10:25.242 "data_offset": 0, 00:10:25.242 "data_size": 65536 00:10:25.242 } 00:10:25.242 ] 00:10:25.242 } 00:10:25.242 } 00:10:25.242 }' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:25.242 BaseBdev2 00:10:25.242 BaseBdev3 00:10:25.242 BaseBdev4' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.242 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.501 [2024-09-30 14:09:29.901984] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.501 "name": "Existed_Raid", 00:10:25.501 "uuid": "7a4a2b40-ea7e-41b0-be60-8bea30f80cec", 00:10:25.501 "strip_size_kb": 0, 00:10:25.501 "state": "online", 00:10:25.501 "raid_level": "raid1", 00:10:25.501 "superblock": false, 00:10:25.501 "num_base_bdevs": 4, 00:10:25.501 "num_base_bdevs_discovered": 3, 00:10:25.501 "num_base_bdevs_operational": 3, 00:10:25.501 "base_bdevs_list": [ 00:10:25.501 { 00:10:25.501 "name": null, 00:10:25.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.501 "is_configured": false, 00:10:25.501 "data_offset": 0, 00:10:25.501 "data_size": 65536 00:10:25.501 }, 00:10:25.501 { 00:10:25.501 "name": "BaseBdev2", 00:10:25.501 "uuid": "5051e183-8dcc-43d7-8e84-ce0a5014ecf3", 00:10:25.501 "is_configured": true, 00:10:25.501 "data_offset": 0, 00:10:25.501 "data_size": 65536 00:10:25.501 }, 00:10:25.501 { 00:10:25.501 "name": "BaseBdev3", 00:10:25.501 "uuid": "42343bfa-d180-44d6-92b4-ccb0a97395e7", 00:10:25.501 "is_configured": true, 00:10:25.501 "data_offset": 0, 00:10:25.501 "data_size": 65536 00:10:25.501 }, 00:10:25.501 { 00:10:25.501 "name": "BaseBdev4", 00:10:25.501 "uuid": "0edb7ad7-3bca-44fe-8026-1e1be8c524d7", 00:10:25.501 "is_configured": true, 00:10:25.501 "data_offset": 0, 00:10:25.501 "data_size": 65536 00:10:25.501 } 00:10:25.501 ] 00:10:25.501 }' 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.501 14:09:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.761 [2024-09-30 14:09:30.364287] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.761 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.761 [2024-09-30 14:09:30.411063] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 [2024-09-30 14:09:30.481985] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:26.021 [2024-09-30 14:09:30.482078] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:26.021 [2024-09-30 14:09:30.493228] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:26.021 [2024-09-30 14:09:30.493288] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:26.021 [2024-09-30 14:09:30.493304] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 BaseBdev2 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.021 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.021 [ 00:10:26.021 { 00:10:26.021 "name": "BaseBdev2", 00:10:26.021 "aliases": [ 00:10:26.021 "79234fda-f857-4dcb-8f87-bab71d53d7e5" 00:10:26.021 ], 00:10:26.021 "product_name": "Malloc disk", 00:10:26.021 "block_size": 512, 00:10:26.021 "num_blocks": 65536, 00:10:26.021 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:26.021 "assigned_rate_limits": { 00:10:26.021 "rw_ios_per_sec": 0, 00:10:26.021 "rw_mbytes_per_sec": 0, 00:10:26.022 "r_mbytes_per_sec": 0, 00:10:26.022 "w_mbytes_per_sec": 0 00:10:26.022 }, 00:10:26.022 "claimed": false, 00:10:26.022 "zoned": false, 00:10:26.022 "supported_io_types": { 00:10:26.022 "read": true, 00:10:26.022 "write": true, 00:10:26.022 "unmap": true, 00:10:26.022 "flush": true, 00:10:26.022 "reset": true, 00:10:26.022 "nvme_admin": false, 00:10:26.022 "nvme_io": false, 00:10:26.022 "nvme_io_md": false, 00:10:26.022 "write_zeroes": true, 00:10:26.022 "zcopy": true, 00:10:26.022 "get_zone_info": false, 00:10:26.022 "zone_management": false, 00:10:26.022 "zone_append": false, 00:10:26.022 "compare": false, 00:10:26.022 "compare_and_write": false, 00:10:26.022 "abort": true, 00:10:26.022 "seek_hole": false, 00:10:26.022 "seek_data": false, 00:10:26.022 "copy": true, 00:10:26.022 "nvme_iov_md": false 00:10:26.022 }, 00:10:26.022 "memory_domains": [ 00:10:26.022 { 00:10:26.022 "dma_device_id": "system", 00:10:26.022 "dma_device_type": 1 00:10:26.022 }, 00:10:26.022 { 00:10:26.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.022 "dma_device_type": 2 00:10:26.022 } 00:10:26.022 ], 00:10:26.022 "driver_specific": {} 00:10:26.022 } 00:10:26.022 ] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.022 BaseBdev3 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.022 [ 00:10:26.022 { 00:10:26.022 "name": "BaseBdev3", 00:10:26.022 "aliases": [ 00:10:26.022 "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14" 00:10:26.022 ], 00:10:26.022 "product_name": "Malloc disk", 00:10:26.022 "block_size": 512, 00:10:26.022 "num_blocks": 65536, 00:10:26.022 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:26.022 "assigned_rate_limits": { 00:10:26.022 "rw_ios_per_sec": 0, 00:10:26.022 "rw_mbytes_per_sec": 0, 00:10:26.022 "r_mbytes_per_sec": 0, 00:10:26.022 "w_mbytes_per_sec": 0 00:10:26.022 }, 00:10:26.022 "claimed": false, 00:10:26.022 "zoned": false, 00:10:26.022 "supported_io_types": { 00:10:26.022 "read": true, 00:10:26.022 "write": true, 00:10:26.022 "unmap": true, 00:10:26.022 "flush": true, 00:10:26.022 "reset": true, 00:10:26.022 "nvme_admin": false, 00:10:26.022 "nvme_io": false, 00:10:26.022 "nvme_io_md": false, 00:10:26.022 "write_zeroes": true, 00:10:26.022 "zcopy": true, 00:10:26.022 "get_zone_info": false, 00:10:26.022 "zone_management": false, 00:10:26.022 "zone_append": false, 00:10:26.022 "compare": false, 00:10:26.022 "compare_and_write": false, 00:10:26.022 "abort": true, 00:10:26.022 "seek_hole": false, 00:10:26.022 "seek_data": false, 00:10:26.022 "copy": true, 00:10:26.022 "nvme_iov_md": false 00:10:26.022 }, 00:10:26.022 "memory_domains": [ 00:10:26.022 { 00:10:26.022 "dma_device_id": "system", 00:10:26.022 "dma_device_type": 1 00:10:26.022 }, 00:10:26.022 { 00:10:26.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.022 "dma_device_type": 2 00:10:26.022 } 00:10:26.022 ], 00:10:26.022 "driver_specific": {} 00:10:26.022 } 00:10:26.022 ] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.022 BaseBdev4 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.022 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.282 [ 00:10:26.282 { 00:10:26.282 "name": "BaseBdev4", 00:10:26.282 "aliases": [ 00:10:26.282 "da23e007-dc38-476e-b078-eecde7662ab4" 00:10:26.282 ], 00:10:26.282 "product_name": "Malloc disk", 00:10:26.282 "block_size": 512, 00:10:26.282 "num_blocks": 65536, 00:10:26.282 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:26.282 "assigned_rate_limits": { 00:10:26.282 "rw_ios_per_sec": 0, 00:10:26.282 "rw_mbytes_per_sec": 0, 00:10:26.282 "r_mbytes_per_sec": 0, 00:10:26.282 "w_mbytes_per_sec": 0 00:10:26.282 }, 00:10:26.282 "claimed": false, 00:10:26.282 "zoned": false, 00:10:26.282 "supported_io_types": { 00:10:26.282 "read": true, 00:10:26.282 "write": true, 00:10:26.282 "unmap": true, 00:10:26.282 "flush": true, 00:10:26.282 "reset": true, 00:10:26.282 "nvme_admin": false, 00:10:26.282 "nvme_io": false, 00:10:26.282 "nvme_io_md": false, 00:10:26.282 "write_zeroes": true, 00:10:26.282 "zcopy": true, 00:10:26.282 "get_zone_info": false, 00:10:26.282 "zone_management": false, 00:10:26.282 "zone_append": false, 00:10:26.282 "compare": false, 00:10:26.282 "compare_and_write": false, 00:10:26.282 "abort": true, 00:10:26.282 "seek_hole": false, 00:10:26.282 "seek_data": false, 00:10:26.282 "copy": true, 00:10:26.282 "nvme_iov_md": false 00:10:26.282 }, 00:10:26.282 "memory_domains": [ 00:10:26.282 { 00:10:26.282 "dma_device_id": "system", 00:10:26.282 "dma_device_type": 1 00:10:26.282 }, 00:10:26.282 { 00:10:26.282 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.282 "dma_device_type": 2 00:10:26.282 } 00:10:26.282 ], 00:10:26.282 "driver_specific": {} 00:10:26.282 } 00:10:26.282 ] 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.282 [2024-09-30 14:09:30.708831] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:26.282 [2024-09-30 14:09:30.708886] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:26.282 [2024-09-30 14:09:30.708922] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:26.282 [2024-09-30 14:09:30.710647] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:26.282 [2024-09-30 14:09:30.710694] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.282 "name": "Existed_Raid", 00:10:26.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.282 "strip_size_kb": 0, 00:10:26.282 "state": "configuring", 00:10:26.282 "raid_level": "raid1", 00:10:26.282 "superblock": false, 00:10:26.282 "num_base_bdevs": 4, 00:10:26.282 "num_base_bdevs_discovered": 3, 00:10:26.282 "num_base_bdevs_operational": 4, 00:10:26.282 "base_bdevs_list": [ 00:10:26.282 { 00:10:26.282 "name": "BaseBdev1", 00:10:26.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.282 "is_configured": false, 00:10:26.282 "data_offset": 0, 00:10:26.282 "data_size": 0 00:10:26.282 }, 00:10:26.282 { 00:10:26.282 "name": "BaseBdev2", 00:10:26.282 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:26.282 "is_configured": true, 00:10:26.282 "data_offset": 0, 00:10:26.282 "data_size": 65536 00:10:26.282 }, 00:10:26.282 { 00:10:26.282 "name": "BaseBdev3", 00:10:26.282 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:26.282 "is_configured": true, 00:10:26.282 "data_offset": 0, 00:10:26.282 "data_size": 65536 00:10:26.282 }, 00:10:26.282 { 00:10:26.282 "name": "BaseBdev4", 00:10:26.282 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:26.282 "is_configured": true, 00:10:26.282 "data_offset": 0, 00:10:26.282 "data_size": 65536 00:10:26.282 } 00:10:26.282 ] 00:10:26.282 }' 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.282 14:09:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.541 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:26.541 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.541 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.800 [2024-09-30 14:09:31.199984] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.800 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.801 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.801 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.801 "name": "Existed_Raid", 00:10:26.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.801 "strip_size_kb": 0, 00:10:26.801 "state": "configuring", 00:10:26.801 "raid_level": "raid1", 00:10:26.801 "superblock": false, 00:10:26.801 "num_base_bdevs": 4, 00:10:26.801 "num_base_bdevs_discovered": 2, 00:10:26.801 "num_base_bdevs_operational": 4, 00:10:26.801 "base_bdevs_list": [ 00:10:26.801 { 00:10:26.801 "name": "BaseBdev1", 00:10:26.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.801 "is_configured": false, 00:10:26.801 "data_offset": 0, 00:10:26.801 "data_size": 0 00:10:26.801 }, 00:10:26.801 { 00:10:26.801 "name": null, 00:10:26.801 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:26.801 "is_configured": false, 00:10:26.801 "data_offset": 0, 00:10:26.801 "data_size": 65536 00:10:26.801 }, 00:10:26.801 { 00:10:26.801 "name": "BaseBdev3", 00:10:26.801 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:26.801 "is_configured": true, 00:10:26.801 "data_offset": 0, 00:10:26.801 "data_size": 65536 00:10:26.801 }, 00:10:26.801 { 00:10:26.801 "name": "BaseBdev4", 00:10:26.801 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:26.801 "is_configured": true, 00:10:26.801 "data_offset": 0, 00:10:26.801 "data_size": 65536 00:10:26.801 } 00:10:26.801 ] 00:10:26.801 }' 00:10:26.801 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.801 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 [2024-09-30 14:09:31.618082] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:27.060 BaseBdev1 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 [ 00:10:27.060 { 00:10:27.060 "name": "BaseBdev1", 00:10:27.060 "aliases": [ 00:10:27.060 "ac21385f-c930-4e69-8b87-f8c5f5c98b1f" 00:10:27.060 ], 00:10:27.060 "product_name": "Malloc disk", 00:10:27.060 "block_size": 512, 00:10:27.060 "num_blocks": 65536, 00:10:27.060 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:27.060 "assigned_rate_limits": { 00:10:27.060 "rw_ios_per_sec": 0, 00:10:27.060 "rw_mbytes_per_sec": 0, 00:10:27.060 "r_mbytes_per_sec": 0, 00:10:27.060 "w_mbytes_per_sec": 0 00:10:27.060 }, 00:10:27.060 "claimed": true, 00:10:27.060 "claim_type": "exclusive_write", 00:10:27.060 "zoned": false, 00:10:27.060 "supported_io_types": { 00:10:27.060 "read": true, 00:10:27.060 "write": true, 00:10:27.060 "unmap": true, 00:10:27.060 "flush": true, 00:10:27.060 "reset": true, 00:10:27.060 "nvme_admin": false, 00:10:27.060 "nvme_io": false, 00:10:27.060 "nvme_io_md": false, 00:10:27.060 "write_zeroes": true, 00:10:27.060 "zcopy": true, 00:10:27.060 "get_zone_info": false, 00:10:27.060 "zone_management": false, 00:10:27.060 "zone_append": false, 00:10:27.060 "compare": false, 00:10:27.060 "compare_and_write": false, 00:10:27.060 "abort": true, 00:10:27.060 "seek_hole": false, 00:10:27.060 "seek_data": false, 00:10:27.060 "copy": true, 00:10:27.060 "nvme_iov_md": false 00:10:27.060 }, 00:10:27.060 "memory_domains": [ 00:10:27.060 { 00:10:27.060 "dma_device_id": "system", 00:10:27.060 "dma_device_type": 1 00:10:27.060 }, 00:10:27.060 { 00:10:27.060 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.060 "dma_device_type": 2 00:10:27.060 } 00:10:27.060 ], 00:10:27.060 "driver_specific": {} 00:10:27.060 } 00:10:27.060 ] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.060 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.060 "name": "Existed_Raid", 00:10:27.060 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.060 "strip_size_kb": 0, 00:10:27.060 "state": "configuring", 00:10:27.060 "raid_level": "raid1", 00:10:27.060 "superblock": false, 00:10:27.060 "num_base_bdevs": 4, 00:10:27.060 "num_base_bdevs_discovered": 3, 00:10:27.060 "num_base_bdevs_operational": 4, 00:10:27.060 "base_bdevs_list": [ 00:10:27.060 { 00:10:27.060 "name": "BaseBdev1", 00:10:27.060 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:27.060 "is_configured": true, 00:10:27.060 "data_offset": 0, 00:10:27.060 "data_size": 65536 00:10:27.060 }, 00:10:27.060 { 00:10:27.060 "name": null, 00:10:27.060 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:27.060 "is_configured": false, 00:10:27.060 "data_offset": 0, 00:10:27.060 "data_size": 65536 00:10:27.060 }, 00:10:27.060 { 00:10:27.060 "name": "BaseBdev3", 00:10:27.060 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:27.060 "is_configured": true, 00:10:27.060 "data_offset": 0, 00:10:27.060 "data_size": 65536 00:10:27.060 }, 00:10:27.060 { 00:10:27.060 "name": "BaseBdev4", 00:10:27.060 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:27.060 "is_configured": true, 00:10:27.061 "data_offset": 0, 00:10:27.061 "data_size": 65536 00:10:27.061 } 00:10:27.061 ] 00:10:27.061 }' 00:10:27.061 14:09:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.061 14:09:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.627 [2024-09-30 14:09:32.121247] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.627 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.628 "name": "Existed_Raid", 00:10:27.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.628 "strip_size_kb": 0, 00:10:27.628 "state": "configuring", 00:10:27.628 "raid_level": "raid1", 00:10:27.628 "superblock": false, 00:10:27.628 "num_base_bdevs": 4, 00:10:27.628 "num_base_bdevs_discovered": 2, 00:10:27.628 "num_base_bdevs_operational": 4, 00:10:27.628 "base_bdevs_list": [ 00:10:27.628 { 00:10:27.628 "name": "BaseBdev1", 00:10:27.628 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:27.628 "is_configured": true, 00:10:27.628 "data_offset": 0, 00:10:27.628 "data_size": 65536 00:10:27.628 }, 00:10:27.628 { 00:10:27.628 "name": null, 00:10:27.628 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:27.628 "is_configured": false, 00:10:27.628 "data_offset": 0, 00:10:27.628 "data_size": 65536 00:10:27.628 }, 00:10:27.628 { 00:10:27.628 "name": null, 00:10:27.628 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:27.628 "is_configured": false, 00:10:27.628 "data_offset": 0, 00:10:27.628 "data_size": 65536 00:10:27.628 }, 00:10:27.628 { 00:10:27.628 "name": "BaseBdev4", 00:10:27.628 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:27.628 "is_configured": true, 00:10:27.628 "data_offset": 0, 00:10:27.628 "data_size": 65536 00:10:27.628 } 00:10:27.628 ] 00:10:27.628 }' 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.628 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.197 [2024-09-30 14:09:32.644370] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.197 "name": "Existed_Raid", 00:10:28.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.197 "strip_size_kb": 0, 00:10:28.197 "state": "configuring", 00:10:28.197 "raid_level": "raid1", 00:10:28.197 "superblock": false, 00:10:28.197 "num_base_bdevs": 4, 00:10:28.197 "num_base_bdevs_discovered": 3, 00:10:28.197 "num_base_bdevs_operational": 4, 00:10:28.197 "base_bdevs_list": [ 00:10:28.197 { 00:10:28.197 "name": "BaseBdev1", 00:10:28.197 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:28.197 "is_configured": true, 00:10:28.197 "data_offset": 0, 00:10:28.197 "data_size": 65536 00:10:28.197 }, 00:10:28.197 { 00:10:28.197 "name": null, 00:10:28.197 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:28.197 "is_configured": false, 00:10:28.197 "data_offset": 0, 00:10:28.197 "data_size": 65536 00:10:28.197 }, 00:10:28.197 { 00:10:28.197 "name": "BaseBdev3", 00:10:28.197 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:28.197 "is_configured": true, 00:10:28.197 "data_offset": 0, 00:10:28.197 "data_size": 65536 00:10:28.197 }, 00:10:28.197 { 00:10:28.197 "name": "BaseBdev4", 00:10:28.197 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:28.197 "is_configured": true, 00:10:28.197 "data_offset": 0, 00:10:28.197 "data_size": 65536 00:10:28.197 } 00:10:28.197 ] 00:10:28.197 }' 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.197 14:09:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.764 [2024-09-30 14:09:33.171530] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.764 "name": "Existed_Raid", 00:10:28.764 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.764 "strip_size_kb": 0, 00:10:28.764 "state": "configuring", 00:10:28.764 "raid_level": "raid1", 00:10:28.764 "superblock": false, 00:10:28.764 "num_base_bdevs": 4, 00:10:28.764 "num_base_bdevs_discovered": 2, 00:10:28.764 "num_base_bdevs_operational": 4, 00:10:28.764 "base_bdevs_list": [ 00:10:28.764 { 00:10:28.764 "name": null, 00:10:28.764 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:28.764 "is_configured": false, 00:10:28.764 "data_offset": 0, 00:10:28.764 "data_size": 65536 00:10:28.764 }, 00:10:28.764 { 00:10:28.764 "name": null, 00:10:28.764 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:28.764 "is_configured": false, 00:10:28.764 "data_offset": 0, 00:10:28.764 "data_size": 65536 00:10:28.764 }, 00:10:28.764 { 00:10:28.764 "name": "BaseBdev3", 00:10:28.764 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:28.764 "is_configured": true, 00:10:28.764 "data_offset": 0, 00:10:28.764 "data_size": 65536 00:10:28.764 }, 00:10:28.764 { 00:10:28.764 "name": "BaseBdev4", 00:10:28.764 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:28.764 "is_configured": true, 00:10:28.764 "data_offset": 0, 00:10:28.764 "data_size": 65536 00:10:28.764 } 00:10:28.764 ] 00:10:28.764 }' 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.764 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:29.023 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.024 [2024-09-30 14:09:33.668934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.024 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.282 "name": "Existed_Raid", 00:10:29.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.282 "strip_size_kb": 0, 00:10:29.282 "state": "configuring", 00:10:29.282 "raid_level": "raid1", 00:10:29.282 "superblock": false, 00:10:29.282 "num_base_bdevs": 4, 00:10:29.282 "num_base_bdevs_discovered": 3, 00:10:29.282 "num_base_bdevs_operational": 4, 00:10:29.282 "base_bdevs_list": [ 00:10:29.282 { 00:10:29.282 "name": null, 00:10:29.282 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:29.282 "is_configured": false, 00:10:29.282 "data_offset": 0, 00:10:29.282 "data_size": 65536 00:10:29.282 }, 00:10:29.282 { 00:10:29.282 "name": "BaseBdev2", 00:10:29.282 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:29.282 "is_configured": true, 00:10:29.282 "data_offset": 0, 00:10:29.282 "data_size": 65536 00:10:29.282 }, 00:10:29.282 { 00:10:29.282 "name": "BaseBdev3", 00:10:29.282 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:29.282 "is_configured": true, 00:10:29.282 "data_offset": 0, 00:10:29.282 "data_size": 65536 00:10:29.282 }, 00:10:29.282 { 00:10:29.282 "name": "BaseBdev4", 00:10:29.282 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:29.282 "is_configured": true, 00:10:29.282 "data_offset": 0, 00:10:29.282 "data_size": 65536 00:10:29.282 } 00:10:29.282 ] 00:10:29.282 }' 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.282 14:09:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:29.541 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ac21385f-c930-4e69-8b87-f8c5f5c98b1f 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.800 [2024-09-30 14:09:34.238704] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:29.800 [2024-09-30 14:09:34.238761] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:29.800 [2024-09-30 14:09:34.238770] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:29.800 [2024-09-30 14:09:34.239055] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:10:29.800 [2024-09-30 14:09:34.239181] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:29.800 [2024-09-30 14:09:34.239198] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:29.800 [2024-09-30 14:09:34.239371] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:29.800 NewBaseBdev 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.800 [ 00:10:29.800 { 00:10:29.800 "name": "NewBaseBdev", 00:10:29.800 "aliases": [ 00:10:29.800 "ac21385f-c930-4e69-8b87-f8c5f5c98b1f" 00:10:29.800 ], 00:10:29.800 "product_name": "Malloc disk", 00:10:29.800 "block_size": 512, 00:10:29.800 "num_blocks": 65536, 00:10:29.800 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:29.800 "assigned_rate_limits": { 00:10:29.800 "rw_ios_per_sec": 0, 00:10:29.800 "rw_mbytes_per_sec": 0, 00:10:29.800 "r_mbytes_per_sec": 0, 00:10:29.800 "w_mbytes_per_sec": 0 00:10:29.800 }, 00:10:29.800 "claimed": true, 00:10:29.800 "claim_type": "exclusive_write", 00:10:29.800 "zoned": false, 00:10:29.800 "supported_io_types": { 00:10:29.800 "read": true, 00:10:29.800 "write": true, 00:10:29.800 "unmap": true, 00:10:29.800 "flush": true, 00:10:29.800 "reset": true, 00:10:29.800 "nvme_admin": false, 00:10:29.800 "nvme_io": false, 00:10:29.800 "nvme_io_md": false, 00:10:29.800 "write_zeroes": true, 00:10:29.800 "zcopy": true, 00:10:29.800 "get_zone_info": false, 00:10:29.800 "zone_management": false, 00:10:29.800 "zone_append": false, 00:10:29.800 "compare": false, 00:10:29.800 "compare_and_write": false, 00:10:29.800 "abort": true, 00:10:29.800 "seek_hole": false, 00:10:29.800 "seek_data": false, 00:10:29.800 "copy": true, 00:10:29.800 "nvme_iov_md": false 00:10:29.800 }, 00:10:29.800 "memory_domains": [ 00:10:29.800 { 00:10:29.800 "dma_device_id": "system", 00:10:29.800 "dma_device_type": 1 00:10:29.800 }, 00:10:29.800 { 00:10:29.800 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:29.800 "dma_device_type": 2 00:10:29.800 } 00:10:29.800 ], 00:10:29.800 "driver_specific": {} 00:10:29.800 } 00:10:29.800 ] 00:10:29.800 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.801 "name": "Existed_Raid", 00:10:29.801 "uuid": "3693d9ad-6153-42fc-9d81-e33e57be6044", 00:10:29.801 "strip_size_kb": 0, 00:10:29.801 "state": "online", 00:10:29.801 "raid_level": "raid1", 00:10:29.801 "superblock": false, 00:10:29.801 "num_base_bdevs": 4, 00:10:29.801 "num_base_bdevs_discovered": 4, 00:10:29.801 "num_base_bdevs_operational": 4, 00:10:29.801 "base_bdevs_list": [ 00:10:29.801 { 00:10:29.801 "name": "NewBaseBdev", 00:10:29.801 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:29.801 "is_configured": true, 00:10:29.801 "data_offset": 0, 00:10:29.801 "data_size": 65536 00:10:29.801 }, 00:10:29.801 { 00:10:29.801 "name": "BaseBdev2", 00:10:29.801 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:29.801 "is_configured": true, 00:10:29.801 "data_offset": 0, 00:10:29.801 "data_size": 65536 00:10:29.801 }, 00:10:29.801 { 00:10:29.801 "name": "BaseBdev3", 00:10:29.801 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:29.801 "is_configured": true, 00:10:29.801 "data_offset": 0, 00:10:29.801 "data_size": 65536 00:10:29.801 }, 00:10:29.801 { 00:10:29.801 "name": "BaseBdev4", 00:10:29.801 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:29.801 "is_configured": true, 00:10:29.801 "data_offset": 0, 00:10:29.801 "data_size": 65536 00:10:29.801 } 00:10:29.801 ] 00:10:29.801 }' 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.801 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.060 [2024-09-30 14:09:34.674287] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.060 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:30.060 "name": "Existed_Raid", 00:10:30.060 "aliases": [ 00:10:30.060 "3693d9ad-6153-42fc-9d81-e33e57be6044" 00:10:30.060 ], 00:10:30.060 "product_name": "Raid Volume", 00:10:30.060 "block_size": 512, 00:10:30.060 "num_blocks": 65536, 00:10:30.060 "uuid": "3693d9ad-6153-42fc-9d81-e33e57be6044", 00:10:30.060 "assigned_rate_limits": { 00:10:30.060 "rw_ios_per_sec": 0, 00:10:30.060 "rw_mbytes_per_sec": 0, 00:10:30.060 "r_mbytes_per_sec": 0, 00:10:30.060 "w_mbytes_per_sec": 0 00:10:30.060 }, 00:10:30.060 "claimed": false, 00:10:30.060 "zoned": false, 00:10:30.060 "supported_io_types": { 00:10:30.060 "read": true, 00:10:30.060 "write": true, 00:10:30.060 "unmap": false, 00:10:30.060 "flush": false, 00:10:30.060 "reset": true, 00:10:30.060 "nvme_admin": false, 00:10:30.060 "nvme_io": false, 00:10:30.060 "nvme_io_md": false, 00:10:30.060 "write_zeroes": true, 00:10:30.060 "zcopy": false, 00:10:30.060 "get_zone_info": false, 00:10:30.060 "zone_management": false, 00:10:30.060 "zone_append": false, 00:10:30.060 "compare": false, 00:10:30.060 "compare_and_write": false, 00:10:30.060 "abort": false, 00:10:30.060 "seek_hole": false, 00:10:30.060 "seek_data": false, 00:10:30.060 "copy": false, 00:10:30.060 "nvme_iov_md": false 00:10:30.060 }, 00:10:30.060 "memory_domains": [ 00:10:30.061 { 00:10:30.061 "dma_device_id": "system", 00:10:30.061 "dma_device_type": 1 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.061 "dma_device_type": 2 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "system", 00:10:30.061 "dma_device_type": 1 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.061 "dma_device_type": 2 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "system", 00:10:30.061 "dma_device_type": 1 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.061 "dma_device_type": 2 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "system", 00:10:30.061 "dma_device_type": 1 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.061 "dma_device_type": 2 00:10:30.061 } 00:10:30.061 ], 00:10:30.061 "driver_specific": { 00:10:30.061 "raid": { 00:10:30.061 "uuid": "3693d9ad-6153-42fc-9d81-e33e57be6044", 00:10:30.061 "strip_size_kb": 0, 00:10:30.061 "state": "online", 00:10:30.061 "raid_level": "raid1", 00:10:30.061 "superblock": false, 00:10:30.061 "num_base_bdevs": 4, 00:10:30.061 "num_base_bdevs_discovered": 4, 00:10:30.061 "num_base_bdevs_operational": 4, 00:10:30.061 "base_bdevs_list": [ 00:10:30.061 { 00:10:30.061 "name": "NewBaseBdev", 00:10:30.061 "uuid": "ac21385f-c930-4e69-8b87-f8c5f5c98b1f", 00:10:30.061 "is_configured": true, 00:10:30.061 "data_offset": 0, 00:10:30.061 "data_size": 65536 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "name": "BaseBdev2", 00:10:30.061 "uuid": "79234fda-f857-4dcb-8f87-bab71d53d7e5", 00:10:30.061 "is_configured": true, 00:10:30.061 "data_offset": 0, 00:10:30.061 "data_size": 65536 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "name": "BaseBdev3", 00:10:30.061 "uuid": "ab3ec8e4-abc1-44cd-a4c7-dfe6f80a9c14", 00:10:30.061 "is_configured": true, 00:10:30.061 "data_offset": 0, 00:10:30.061 "data_size": 65536 00:10:30.061 }, 00:10:30.061 { 00:10:30.061 "name": "BaseBdev4", 00:10:30.061 "uuid": "da23e007-dc38-476e-b078-eecde7662ab4", 00:10:30.061 "is_configured": true, 00:10:30.061 "data_offset": 0, 00:10:30.061 "data_size": 65536 00:10:30.061 } 00:10:30.061 ] 00:10:30.061 } 00:10:30.061 } 00:10:30.061 }' 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:30.320 BaseBdev2 00:10:30.320 BaseBdev3 00:10:30.320 BaseBdev4' 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.320 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.321 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.580 [2024-09-30 14:09:34.981463] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.580 [2024-09-30 14:09:34.981500] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:30.580 [2024-09-30 14:09:34.981577] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:30.580 [2024-09-30 14:09:34.981833] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:30.580 [2024-09-30 14:09:34.981850] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 85008 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 85008 ']' 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 85008 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:30.580 14:09:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85008 00:10:30.580 killing process with pid 85008 00:10:30.580 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:30.581 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:30.581 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85008' 00:10:30.581 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 85008 00:10:30.581 [2024-09-30 14:09:35.029182] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:30.581 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 85008 00:10:30.581 [2024-09-30 14:09:35.067951] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:30.840 00:10:30.840 real 0m9.380s 00:10:30.840 user 0m15.951s 00:10:30.840 sys 0m2.037s 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.840 ************************************ 00:10:30.840 END TEST raid_state_function_test 00:10:30.840 ************************************ 00:10:30.840 14:09:35 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:10:30.840 14:09:35 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:30.840 14:09:35 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:30.840 14:09:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:30.840 ************************************ 00:10:30.840 START TEST raid_state_function_test_sb 00:10:30.840 ************************************ 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 true 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=85653 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 85653' 00:10:30.840 Process raid pid: 85653 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 85653 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 85653 ']' 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:30.840 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:30.840 14:09:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.100 [2024-09-30 14:09:35.498675] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:31.100 [2024-09-30 14:09:35.498817] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:31.100 [2024-09-30 14:09:35.636562] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:31.100 [2024-09-30 14:09:35.666472] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:31.100 [2024-09-30 14:09:35.711026] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:31.100 [2024-09-30 14:09:35.752034] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:31.100 [2024-09-30 14:09:35.752068] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.039 [2024-09-30 14:09:36.348965] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:32.039 [2024-09-30 14:09:36.349020] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:32.039 [2024-09-30 14:09:36.349032] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:32.039 [2024-09-30 14:09:36.349055] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:32.039 [2024-09-30 14:09:36.349066] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:32.039 [2024-09-30 14:09:36.349074] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:32.039 [2024-09-30 14:09:36.349081] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:32.039 [2024-09-30 14:09:36.349087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.039 "name": "Existed_Raid", 00:10:32.039 "uuid": "bfb4e5f6-c293-4c4c-b142-50b224835ecd", 00:10:32.039 "strip_size_kb": 0, 00:10:32.039 "state": "configuring", 00:10:32.039 "raid_level": "raid1", 00:10:32.039 "superblock": true, 00:10:32.039 "num_base_bdevs": 4, 00:10:32.039 "num_base_bdevs_discovered": 0, 00:10:32.039 "num_base_bdevs_operational": 4, 00:10:32.039 "base_bdevs_list": [ 00:10:32.039 { 00:10:32.039 "name": "BaseBdev1", 00:10:32.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.039 "is_configured": false, 00:10:32.039 "data_offset": 0, 00:10:32.039 "data_size": 0 00:10:32.039 }, 00:10:32.039 { 00:10:32.039 "name": "BaseBdev2", 00:10:32.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.039 "is_configured": false, 00:10:32.039 "data_offset": 0, 00:10:32.039 "data_size": 0 00:10:32.039 }, 00:10:32.039 { 00:10:32.039 "name": "BaseBdev3", 00:10:32.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.039 "is_configured": false, 00:10:32.039 "data_offset": 0, 00:10:32.039 "data_size": 0 00:10:32.039 }, 00:10:32.039 { 00:10:32.039 "name": "BaseBdev4", 00:10:32.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.039 "is_configured": false, 00:10:32.039 "data_offset": 0, 00:10:32.039 "data_size": 0 00:10:32.039 } 00:10:32.039 ] 00:10:32.039 }' 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.039 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 [2024-09-30 14:09:36.792109] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:32.300 [2024-09-30 14:09:36.792151] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 [2024-09-30 14:09:36.804109] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:32.300 [2024-09-30 14:09:36.804148] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:32.300 [2024-09-30 14:09:36.804158] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:32.300 [2024-09-30 14:09:36.804165] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:32.300 [2024-09-30 14:09:36.804174] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:32.300 [2024-09-30 14:09:36.804180] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:32.300 [2024-09-30 14:09:36.804188] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:32.300 [2024-09-30 14:09:36.804196] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 [2024-09-30 14:09:36.824579] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:32.300 BaseBdev1 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 [ 00:10:32.300 { 00:10:32.300 "name": "BaseBdev1", 00:10:32.300 "aliases": [ 00:10:32.300 "0510f119-bfe2-4336-a89c-39bb9ace63e3" 00:10:32.300 ], 00:10:32.300 "product_name": "Malloc disk", 00:10:32.300 "block_size": 512, 00:10:32.300 "num_blocks": 65536, 00:10:32.300 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:32.300 "assigned_rate_limits": { 00:10:32.300 "rw_ios_per_sec": 0, 00:10:32.300 "rw_mbytes_per_sec": 0, 00:10:32.300 "r_mbytes_per_sec": 0, 00:10:32.300 "w_mbytes_per_sec": 0 00:10:32.300 }, 00:10:32.300 "claimed": true, 00:10:32.300 "claim_type": "exclusive_write", 00:10:32.300 "zoned": false, 00:10:32.300 "supported_io_types": { 00:10:32.300 "read": true, 00:10:32.300 "write": true, 00:10:32.300 "unmap": true, 00:10:32.300 "flush": true, 00:10:32.300 "reset": true, 00:10:32.300 "nvme_admin": false, 00:10:32.300 "nvme_io": false, 00:10:32.300 "nvme_io_md": false, 00:10:32.300 "write_zeroes": true, 00:10:32.300 "zcopy": true, 00:10:32.300 "get_zone_info": false, 00:10:32.300 "zone_management": false, 00:10:32.300 "zone_append": false, 00:10:32.300 "compare": false, 00:10:32.300 "compare_and_write": false, 00:10:32.300 "abort": true, 00:10:32.300 "seek_hole": false, 00:10:32.300 "seek_data": false, 00:10:32.300 "copy": true, 00:10:32.300 "nvme_iov_md": false 00:10:32.300 }, 00:10:32.300 "memory_domains": [ 00:10:32.300 { 00:10:32.300 "dma_device_id": "system", 00:10:32.300 "dma_device_type": 1 00:10:32.300 }, 00:10:32.300 { 00:10:32.300 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.300 "dma_device_type": 2 00:10:32.300 } 00:10:32.300 ], 00:10:32.300 "driver_specific": {} 00:10:32.300 } 00:10:32.300 ] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.300 "name": "Existed_Raid", 00:10:32.300 "uuid": "dd7f1e4f-f448-4aaa-8bd4-7dedbe090db1", 00:10:32.300 "strip_size_kb": 0, 00:10:32.300 "state": "configuring", 00:10:32.300 "raid_level": "raid1", 00:10:32.300 "superblock": true, 00:10:32.300 "num_base_bdevs": 4, 00:10:32.300 "num_base_bdevs_discovered": 1, 00:10:32.300 "num_base_bdevs_operational": 4, 00:10:32.300 "base_bdevs_list": [ 00:10:32.300 { 00:10:32.300 "name": "BaseBdev1", 00:10:32.300 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:32.300 "is_configured": true, 00:10:32.300 "data_offset": 2048, 00:10:32.300 "data_size": 63488 00:10:32.300 }, 00:10:32.300 { 00:10:32.300 "name": "BaseBdev2", 00:10:32.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.300 "is_configured": false, 00:10:32.300 "data_offset": 0, 00:10:32.300 "data_size": 0 00:10:32.300 }, 00:10:32.300 { 00:10:32.300 "name": "BaseBdev3", 00:10:32.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.300 "is_configured": false, 00:10:32.300 "data_offset": 0, 00:10:32.300 "data_size": 0 00:10:32.300 }, 00:10:32.300 { 00:10:32.300 "name": "BaseBdev4", 00:10:32.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.300 "is_configured": false, 00:10:32.300 "data_offset": 0, 00:10:32.300 "data_size": 0 00:10:32.300 } 00:10:32.300 ] 00:10:32.300 }' 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.300 14:09:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.870 [2024-09-30 14:09:37.311807] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:32.870 [2024-09-30 14:09:37.311879] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.870 [2024-09-30 14:09:37.319801] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:32.870 [2024-09-30 14:09:37.321609] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:32.870 [2024-09-30 14:09:37.321652] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:32.870 [2024-09-30 14:09:37.321662] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:32.870 [2024-09-30 14:09:37.321671] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:32.870 [2024-09-30 14:09:37.321679] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:32.870 [2024-09-30 14:09:37.321685] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.870 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.871 "name": "Existed_Raid", 00:10:32.871 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:32.871 "strip_size_kb": 0, 00:10:32.871 "state": "configuring", 00:10:32.871 "raid_level": "raid1", 00:10:32.871 "superblock": true, 00:10:32.871 "num_base_bdevs": 4, 00:10:32.871 "num_base_bdevs_discovered": 1, 00:10:32.871 "num_base_bdevs_operational": 4, 00:10:32.871 "base_bdevs_list": [ 00:10:32.871 { 00:10:32.871 "name": "BaseBdev1", 00:10:32.871 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:32.871 "is_configured": true, 00:10:32.871 "data_offset": 2048, 00:10:32.871 "data_size": 63488 00:10:32.871 }, 00:10:32.871 { 00:10:32.871 "name": "BaseBdev2", 00:10:32.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.871 "is_configured": false, 00:10:32.871 "data_offset": 0, 00:10:32.871 "data_size": 0 00:10:32.871 }, 00:10:32.871 { 00:10:32.871 "name": "BaseBdev3", 00:10:32.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.871 "is_configured": false, 00:10:32.871 "data_offset": 0, 00:10:32.871 "data_size": 0 00:10:32.871 }, 00:10:32.871 { 00:10:32.871 "name": "BaseBdev4", 00:10:32.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.871 "is_configured": false, 00:10:32.871 "data_offset": 0, 00:10:32.871 "data_size": 0 00:10:32.871 } 00:10:32.871 ] 00:10:32.871 }' 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.871 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.131 [2024-09-30 14:09:37.723143] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:33.131 BaseBdev2 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.131 [ 00:10:33.131 { 00:10:33.131 "name": "BaseBdev2", 00:10:33.131 "aliases": [ 00:10:33.131 "9ae41f54-ed46-4d33-827b-45a21dbf6f49" 00:10:33.131 ], 00:10:33.131 "product_name": "Malloc disk", 00:10:33.131 "block_size": 512, 00:10:33.131 "num_blocks": 65536, 00:10:33.131 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:33.131 "assigned_rate_limits": { 00:10:33.131 "rw_ios_per_sec": 0, 00:10:33.131 "rw_mbytes_per_sec": 0, 00:10:33.131 "r_mbytes_per_sec": 0, 00:10:33.131 "w_mbytes_per_sec": 0 00:10:33.131 }, 00:10:33.131 "claimed": true, 00:10:33.131 "claim_type": "exclusive_write", 00:10:33.131 "zoned": false, 00:10:33.131 "supported_io_types": { 00:10:33.131 "read": true, 00:10:33.131 "write": true, 00:10:33.131 "unmap": true, 00:10:33.131 "flush": true, 00:10:33.131 "reset": true, 00:10:33.131 "nvme_admin": false, 00:10:33.131 "nvme_io": false, 00:10:33.131 "nvme_io_md": false, 00:10:33.131 "write_zeroes": true, 00:10:33.131 "zcopy": true, 00:10:33.131 "get_zone_info": false, 00:10:33.131 "zone_management": false, 00:10:33.131 "zone_append": false, 00:10:33.131 "compare": false, 00:10:33.131 "compare_and_write": false, 00:10:33.131 "abort": true, 00:10:33.131 "seek_hole": false, 00:10:33.131 "seek_data": false, 00:10:33.131 "copy": true, 00:10:33.131 "nvme_iov_md": false 00:10:33.131 }, 00:10:33.131 "memory_domains": [ 00:10:33.131 { 00:10:33.131 "dma_device_id": "system", 00:10:33.131 "dma_device_type": 1 00:10:33.131 }, 00:10:33.131 { 00:10:33.131 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.131 "dma_device_type": 2 00:10:33.131 } 00:10:33.131 ], 00:10:33.131 "driver_specific": {} 00:10:33.131 } 00:10:33.131 ] 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.131 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.132 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.391 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.391 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.391 "name": "Existed_Raid", 00:10:33.391 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:33.391 "strip_size_kb": 0, 00:10:33.391 "state": "configuring", 00:10:33.391 "raid_level": "raid1", 00:10:33.391 "superblock": true, 00:10:33.391 "num_base_bdevs": 4, 00:10:33.391 "num_base_bdevs_discovered": 2, 00:10:33.391 "num_base_bdevs_operational": 4, 00:10:33.391 "base_bdevs_list": [ 00:10:33.391 { 00:10:33.391 "name": "BaseBdev1", 00:10:33.391 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:33.391 "is_configured": true, 00:10:33.391 "data_offset": 2048, 00:10:33.391 "data_size": 63488 00:10:33.391 }, 00:10:33.391 { 00:10:33.391 "name": "BaseBdev2", 00:10:33.391 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:33.391 "is_configured": true, 00:10:33.391 "data_offset": 2048, 00:10:33.391 "data_size": 63488 00:10:33.391 }, 00:10:33.391 { 00:10:33.391 "name": "BaseBdev3", 00:10:33.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.391 "is_configured": false, 00:10:33.391 "data_offset": 0, 00:10:33.391 "data_size": 0 00:10:33.391 }, 00:10:33.391 { 00:10:33.391 "name": "BaseBdev4", 00:10:33.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.391 "is_configured": false, 00:10:33.391 "data_offset": 0, 00:10:33.391 "data_size": 0 00:10:33.391 } 00:10:33.391 ] 00:10:33.391 }' 00:10:33.391 14:09:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.391 14:09:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.651 [2024-09-30 14:09:38.193197] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:33.651 BaseBdev3 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.651 [ 00:10:33.651 { 00:10:33.651 "name": "BaseBdev3", 00:10:33.651 "aliases": [ 00:10:33.651 "33e4e8d6-1468-422b-a762-ec16a3afa682" 00:10:33.651 ], 00:10:33.651 "product_name": "Malloc disk", 00:10:33.651 "block_size": 512, 00:10:33.651 "num_blocks": 65536, 00:10:33.651 "uuid": "33e4e8d6-1468-422b-a762-ec16a3afa682", 00:10:33.651 "assigned_rate_limits": { 00:10:33.651 "rw_ios_per_sec": 0, 00:10:33.651 "rw_mbytes_per_sec": 0, 00:10:33.651 "r_mbytes_per_sec": 0, 00:10:33.651 "w_mbytes_per_sec": 0 00:10:33.651 }, 00:10:33.651 "claimed": true, 00:10:33.651 "claim_type": "exclusive_write", 00:10:33.651 "zoned": false, 00:10:33.651 "supported_io_types": { 00:10:33.651 "read": true, 00:10:33.651 "write": true, 00:10:33.651 "unmap": true, 00:10:33.651 "flush": true, 00:10:33.651 "reset": true, 00:10:33.651 "nvme_admin": false, 00:10:33.651 "nvme_io": false, 00:10:33.651 "nvme_io_md": false, 00:10:33.651 "write_zeroes": true, 00:10:33.651 "zcopy": true, 00:10:33.651 "get_zone_info": false, 00:10:33.651 "zone_management": false, 00:10:33.651 "zone_append": false, 00:10:33.651 "compare": false, 00:10:33.651 "compare_and_write": false, 00:10:33.651 "abort": true, 00:10:33.651 "seek_hole": false, 00:10:33.651 "seek_data": false, 00:10:33.651 "copy": true, 00:10:33.651 "nvme_iov_md": false 00:10:33.651 }, 00:10:33.651 "memory_domains": [ 00:10:33.651 { 00:10:33.651 "dma_device_id": "system", 00:10:33.651 "dma_device_type": 1 00:10:33.651 }, 00:10:33.651 { 00:10:33.651 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.651 "dma_device_type": 2 00:10:33.651 } 00:10:33.651 ], 00:10:33.651 "driver_specific": {} 00:10:33.651 } 00:10:33.651 ] 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.651 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.652 "name": "Existed_Raid", 00:10:33.652 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:33.652 "strip_size_kb": 0, 00:10:33.652 "state": "configuring", 00:10:33.652 "raid_level": "raid1", 00:10:33.652 "superblock": true, 00:10:33.652 "num_base_bdevs": 4, 00:10:33.652 "num_base_bdevs_discovered": 3, 00:10:33.652 "num_base_bdevs_operational": 4, 00:10:33.652 "base_bdevs_list": [ 00:10:33.652 { 00:10:33.652 "name": "BaseBdev1", 00:10:33.652 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:33.652 "is_configured": true, 00:10:33.652 "data_offset": 2048, 00:10:33.652 "data_size": 63488 00:10:33.652 }, 00:10:33.652 { 00:10:33.652 "name": "BaseBdev2", 00:10:33.652 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:33.652 "is_configured": true, 00:10:33.652 "data_offset": 2048, 00:10:33.652 "data_size": 63488 00:10:33.652 }, 00:10:33.652 { 00:10:33.652 "name": "BaseBdev3", 00:10:33.652 "uuid": "33e4e8d6-1468-422b-a762-ec16a3afa682", 00:10:33.652 "is_configured": true, 00:10:33.652 "data_offset": 2048, 00:10:33.652 "data_size": 63488 00:10:33.652 }, 00:10:33.652 { 00:10:33.652 "name": "BaseBdev4", 00:10:33.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.652 "is_configured": false, 00:10:33.652 "data_offset": 0, 00:10:33.652 "data_size": 0 00:10:33.652 } 00:10:33.652 ] 00:10:33.652 }' 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.652 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.221 [2024-09-30 14:09:38.647292] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:34.221 [2024-09-30 14:09:38.647497] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:34.221 [2024-09-30 14:09:38.647518] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:34.221 [2024-09-30 14:09:38.647800] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:34.221 BaseBdev4 00:10:34.221 [2024-09-30 14:09:38.647975] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:34.221 [2024-09-30 14:09:38.647993] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:34.221 [2024-09-30 14:09:38.648131] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.221 [ 00:10:34.221 { 00:10:34.221 "name": "BaseBdev4", 00:10:34.221 "aliases": [ 00:10:34.221 "a52e536e-880b-4f52-aa39-d8d7a58ab980" 00:10:34.221 ], 00:10:34.221 "product_name": "Malloc disk", 00:10:34.221 "block_size": 512, 00:10:34.221 "num_blocks": 65536, 00:10:34.221 "uuid": "a52e536e-880b-4f52-aa39-d8d7a58ab980", 00:10:34.221 "assigned_rate_limits": { 00:10:34.221 "rw_ios_per_sec": 0, 00:10:34.221 "rw_mbytes_per_sec": 0, 00:10:34.221 "r_mbytes_per_sec": 0, 00:10:34.221 "w_mbytes_per_sec": 0 00:10:34.221 }, 00:10:34.221 "claimed": true, 00:10:34.221 "claim_type": "exclusive_write", 00:10:34.221 "zoned": false, 00:10:34.221 "supported_io_types": { 00:10:34.221 "read": true, 00:10:34.221 "write": true, 00:10:34.221 "unmap": true, 00:10:34.221 "flush": true, 00:10:34.221 "reset": true, 00:10:34.221 "nvme_admin": false, 00:10:34.221 "nvme_io": false, 00:10:34.221 "nvme_io_md": false, 00:10:34.221 "write_zeroes": true, 00:10:34.221 "zcopy": true, 00:10:34.221 "get_zone_info": false, 00:10:34.221 "zone_management": false, 00:10:34.221 "zone_append": false, 00:10:34.221 "compare": false, 00:10:34.221 "compare_and_write": false, 00:10:34.221 "abort": true, 00:10:34.221 "seek_hole": false, 00:10:34.221 "seek_data": false, 00:10:34.221 "copy": true, 00:10:34.221 "nvme_iov_md": false 00:10:34.221 }, 00:10:34.221 "memory_domains": [ 00:10:34.221 { 00:10:34.221 "dma_device_id": "system", 00:10:34.221 "dma_device_type": 1 00:10:34.221 }, 00:10:34.221 { 00:10:34.221 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.221 "dma_device_type": 2 00:10:34.221 } 00:10:34.221 ], 00:10:34.221 "driver_specific": {} 00:10:34.221 } 00:10:34.221 ] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.221 "name": "Existed_Raid", 00:10:34.221 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:34.221 "strip_size_kb": 0, 00:10:34.221 "state": "online", 00:10:34.221 "raid_level": "raid1", 00:10:34.221 "superblock": true, 00:10:34.221 "num_base_bdevs": 4, 00:10:34.221 "num_base_bdevs_discovered": 4, 00:10:34.221 "num_base_bdevs_operational": 4, 00:10:34.221 "base_bdevs_list": [ 00:10:34.221 { 00:10:34.221 "name": "BaseBdev1", 00:10:34.221 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:34.221 "is_configured": true, 00:10:34.221 "data_offset": 2048, 00:10:34.221 "data_size": 63488 00:10:34.221 }, 00:10:34.221 { 00:10:34.221 "name": "BaseBdev2", 00:10:34.221 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:34.221 "is_configured": true, 00:10:34.221 "data_offset": 2048, 00:10:34.221 "data_size": 63488 00:10:34.221 }, 00:10:34.221 { 00:10:34.221 "name": "BaseBdev3", 00:10:34.221 "uuid": "33e4e8d6-1468-422b-a762-ec16a3afa682", 00:10:34.221 "is_configured": true, 00:10:34.221 "data_offset": 2048, 00:10:34.221 "data_size": 63488 00:10:34.221 }, 00:10:34.221 { 00:10:34.221 "name": "BaseBdev4", 00:10:34.221 "uuid": "a52e536e-880b-4f52-aa39-d8d7a58ab980", 00:10:34.221 "is_configured": true, 00:10:34.221 "data_offset": 2048, 00:10:34.221 "data_size": 63488 00:10:34.221 } 00:10:34.221 ] 00:10:34.221 }' 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.221 14:09:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.481 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.481 [2024-09-30 14:09:39.118883] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:34.741 "name": "Existed_Raid", 00:10:34.741 "aliases": [ 00:10:34.741 "cc0d653f-3e6f-43cd-a87d-fb82013b2884" 00:10:34.741 ], 00:10:34.741 "product_name": "Raid Volume", 00:10:34.741 "block_size": 512, 00:10:34.741 "num_blocks": 63488, 00:10:34.741 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:34.741 "assigned_rate_limits": { 00:10:34.741 "rw_ios_per_sec": 0, 00:10:34.741 "rw_mbytes_per_sec": 0, 00:10:34.741 "r_mbytes_per_sec": 0, 00:10:34.741 "w_mbytes_per_sec": 0 00:10:34.741 }, 00:10:34.741 "claimed": false, 00:10:34.741 "zoned": false, 00:10:34.741 "supported_io_types": { 00:10:34.741 "read": true, 00:10:34.741 "write": true, 00:10:34.741 "unmap": false, 00:10:34.741 "flush": false, 00:10:34.741 "reset": true, 00:10:34.741 "nvme_admin": false, 00:10:34.741 "nvme_io": false, 00:10:34.741 "nvme_io_md": false, 00:10:34.741 "write_zeroes": true, 00:10:34.741 "zcopy": false, 00:10:34.741 "get_zone_info": false, 00:10:34.741 "zone_management": false, 00:10:34.741 "zone_append": false, 00:10:34.741 "compare": false, 00:10:34.741 "compare_and_write": false, 00:10:34.741 "abort": false, 00:10:34.741 "seek_hole": false, 00:10:34.741 "seek_data": false, 00:10:34.741 "copy": false, 00:10:34.741 "nvme_iov_md": false 00:10:34.741 }, 00:10:34.741 "memory_domains": [ 00:10:34.741 { 00:10:34.741 "dma_device_id": "system", 00:10:34.741 "dma_device_type": 1 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.741 "dma_device_type": 2 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "system", 00:10:34.741 "dma_device_type": 1 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.741 "dma_device_type": 2 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "system", 00:10:34.741 "dma_device_type": 1 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.741 "dma_device_type": 2 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "system", 00:10:34.741 "dma_device_type": 1 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.741 "dma_device_type": 2 00:10:34.741 } 00:10:34.741 ], 00:10:34.741 "driver_specific": { 00:10:34.741 "raid": { 00:10:34.741 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:34.741 "strip_size_kb": 0, 00:10:34.741 "state": "online", 00:10:34.741 "raid_level": "raid1", 00:10:34.741 "superblock": true, 00:10:34.741 "num_base_bdevs": 4, 00:10:34.741 "num_base_bdevs_discovered": 4, 00:10:34.741 "num_base_bdevs_operational": 4, 00:10:34.741 "base_bdevs_list": [ 00:10:34.741 { 00:10:34.741 "name": "BaseBdev1", 00:10:34.741 "uuid": "0510f119-bfe2-4336-a89c-39bb9ace63e3", 00:10:34.741 "is_configured": true, 00:10:34.741 "data_offset": 2048, 00:10:34.741 "data_size": 63488 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "name": "BaseBdev2", 00:10:34.741 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:34.741 "is_configured": true, 00:10:34.741 "data_offset": 2048, 00:10:34.741 "data_size": 63488 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "name": "BaseBdev3", 00:10:34.741 "uuid": "33e4e8d6-1468-422b-a762-ec16a3afa682", 00:10:34.741 "is_configured": true, 00:10:34.741 "data_offset": 2048, 00:10:34.741 "data_size": 63488 00:10:34.741 }, 00:10:34.741 { 00:10:34.741 "name": "BaseBdev4", 00:10:34.741 "uuid": "a52e536e-880b-4f52-aa39-d8d7a58ab980", 00:10:34.741 "is_configured": true, 00:10:34.741 "data_offset": 2048, 00:10:34.741 "data_size": 63488 00:10:34.741 } 00:10:34.741 ] 00:10:34.741 } 00:10:34.741 } 00:10:34.741 }' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:34.741 BaseBdev2 00:10:34.741 BaseBdev3 00:10:34.741 BaseBdev4' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.741 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:35.001 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.002 [2024-09-30 14:09:39.438068] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.002 "name": "Existed_Raid", 00:10:35.002 "uuid": "cc0d653f-3e6f-43cd-a87d-fb82013b2884", 00:10:35.002 "strip_size_kb": 0, 00:10:35.002 "state": "online", 00:10:35.002 "raid_level": "raid1", 00:10:35.002 "superblock": true, 00:10:35.002 "num_base_bdevs": 4, 00:10:35.002 "num_base_bdevs_discovered": 3, 00:10:35.002 "num_base_bdevs_operational": 3, 00:10:35.002 "base_bdevs_list": [ 00:10:35.002 { 00:10:35.002 "name": null, 00:10:35.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.002 "is_configured": false, 00:10:35.002 "data_offset": 0, 00:10:35.002 "data_size": 63488 00:10:35.002 }, 00:10:35.002 { 00:10:35.002 "name": "BaseBdev2", 00:10:35.002 "uuid": "9ae41f54-ed46-4d33-827b-45a21dbf6f49", 00:10:35.002 "is_configured": true, 00:10:35.002 "data_offset": 2048, 00:10:35.002 "data_size": 63488 00:10:35.002 }, 00:10:35.002 { 00:10:35.002 "name": "BaseBdev3", 00:10:35.002 "uuid": "33e4e8d6-1468-422b-a762-ec16a3afa682", 00:10:35.002 "is_configured": true, 00:10:35.002 "data_offset": 2048, 00:10:35.002 "data_size": 63488 00:10:35.002 }, 00:10:35.002 { 00:10:35.002 "name": "BaseBdev4", 00:10:35.002 "uuid": "a52e536e-880b-4f52-aa39-d8d7a58ab980", 00:10:35.002 "is_configured": true, 00:10:35.002 "data_offset": 2048, 00:10:35.002 "data_size": 63488 00:10:35.002 } 00:10:35.002 ] 00:10:35.002 }' 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.002 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.262 [2024-09-30 14:09:39.896425] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.262 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.523 [2024-09-30 14:09:39.967375] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.523 14:09:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.523 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.523 [2024-09-30 14:09:40.014251] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:35.523 [2024-09-30 14:09:40.014349] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:35.523 [2024-09-30 14:09:40.025671] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:35.523 [2024-09-30 14:09:40.025716] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:35.524 [2024-09-30 14:09:40.025726] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.524 BaseBdev2 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.524 [ 00:10:35.524 { 00:10:35.524 "name": "BaseBdev2", 00:10:35.524 "aliases": [ 00:10:35.524 "3fded517-2e98-4ef5-8e32-97e40e08529d" 00:10:35.524 ], 00:10:35.524 "product_name": "Malloc disk", 00:10:35.524 "block_size": 512, 00:10:35.524 "num_blocks": 65536, 00:10:35.524 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:35.524 "assigned_rate_limits": { 00:10:35.524 "rw_ios_per_sec": 0, 00:10:35.524 "rw_mbytes_per_sec": 0, 00:10:35.524 "r_mbytes_per_sec": 0, 00:10:35.524 "w_mbytes_per_sec": 0 00:10:35.524 }, 00:10:35.524 "claimed": false, 00:10:35.524 "zoned": false, 00:10:35.524 "supported_io_types": { 00:10:35.524 "read": true, 00:10:35.524 "write": true, 00:10:35.524 "unmap": true, 00:10:35.524 "flush": true, 00:10:35.524 "reset": true, 00:10:35.524 "nvme_admin": false, 00:10:35.524 "nvme_io": false, 00:10:35.524 "nvme_io_md": false, 00:10:35.524 "write_zeroes": true, 00:10:35.524 "zcopy": true, 00:10:35.524 "get_zone_info": false, 00:10:35.524 "zone_management": false, 00:10:35.524 "zone_append": false, 00:10:35.524 "compare": false, 00:10:35.524 "compare_and_write": false, 00:10:35.524 "abort": true, 00:10:35.524 "seek_hole": false, 00:10:35.524 "seek_data": false, 00:10:35.524 "copy": true, 00:10:35.524 "nvme_iov_md": false 00:10:35.524 }, 00:10:35.524 "memory_domains": [ 00:10:35.524 { 00:10:35.524 "dma_device_id": "system", 00:10:35.524 "dma_device_type": 1 00:10:35.524 }, 00:10:35.524 { 00:10:35.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.524 "dma_device_type": 2 00:10:35.524 } 00:10:35.524 ], 00:10:35.524 "driver_specific": {} 00:10:35.524 } 00:10:35.524 ] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:35.524 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.525 BaseBdev3 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.525 [ 00:10:35.525 { 00:10:35.525 "name": "BaseBdev3", 00:10:35.525 "aliases": [ 00:10:35.525 "aaaa5729-5484-4673-a639-c0151a44efdd" 00:10:35.525 ], 00:10:35.525 "product_name": "Malloc disk", 00:10:35.525 "block_size": 512, 00:10:35.525 "num_blocks": 65536, 00:10:35.525 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:35.525 "assigned_rate_limits": { 00:10:35.525 "rw_ios_per_sec": 0, 00:10:35.525 "rw_mbytes_per_sec": 0, 00:10:35.525 "r_mbytes_per_sec": 0, 00:10:35.525 "w_mbytes_per_sec": 0 00:10:35.525 }, 00:10:35.525 "claimed": false, 00:10:35.525 "zoned": false, 00:10:35.525 "supported_io_types": { 00:10:35.525 "read": true, 00:10:35.525 "write": true, 00:10:35.525 "unmap": true, 00:10:35.525 "flush": true, 00:10:35.525 "reset": true, 00:10:35.525 "nvme_admin": false, 00:10:35.525 "nvme_io": false, 00:10:35.525 "nvme_io_md": false, 00:10:35.525 "write_zeroes": true, 00:10:35.525 "zcopy": true, 00:10:35.525 "get_zone_info": false, 00:10:35.525 "zone_management": false, 00:10:35.525 "zone_append": false, 00:10:35.525 "compare": false, 00:10:35.525 "compare_and_write": false, 00:10:35.525 "abort": true, 00:10:35.525 "seek_hole": false, 00:10:35.525 "seek_data": false, 00:10:35.525 "copy": true, 00:10:35.525 "nvme_iov_md": false 00:10:35.525 }, 00:10:35.525 "memory_domains": [ 00:10:35.525 { 00:10:35.525 "dma_device_id": "system", 00:10:35.525 "dma_device_type": 1 00:10:35.525 }, 00:10:35.525 { 00:10:35.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.525 "dma_device_type": 2 00:10:35.525 } 00:10:35.525 ], 00:10:35.525 "driver_specific": {} 00:10:35.525 } 00:10:35.525 ] 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.525 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.785 BaseBdev4 00:10:35.785 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.786 [ 00:10:35.786 { 00:10:35.786 "name": "BaseBdev4", 00:10:35.786 "aliases": [ 00:10:35.786 "3af2c9dc-d0c6-4fdb-bd68-65349f398032" 00:10:35.786 ], 00:10:35.786 "product_name": "Malloc disk", 00:10:35.786 "block_size": 512, 00:10:35.786 "num_blocks": 65536, 00:10:35.786 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:35.786 "assigned_rate_limits": { 00:10:35.786 "rw_ios_per_sec": 0, 00:10:35.786 "rw_mbytes_per_sec": 0, 00:10:35.786 "r_mbytes_per_sec": 0, 00:10:35.786 "w_mbytes_per_sec": 0 00:10:35.786 }, 00:10:35.786 "claimed": false, 00:10:35.786 "zoned": false, 00:10:35.786 "supported_io_types": { 00:10:35.786 "read": true, 00:10:35.786 "write": true, 00:10:35.786 "unmap": true, 00:10:35.786 "flush": true, 00:10:35.786 "reset": true, 00:10:35.786 "nvme_admin": false, 00:10:35.786 "nvme_io": false, 00:10:35.786 "nvme_io_md": false, 00:10:35.786 "write_zeroes": true, 00:10:35.786 "zcopy": true, 00:10:35.786 "get_zone_info": false, 00:10:35.786 "zone_management": false, 00:10:35.786 "zone_append": false, 00:10:35.786 "compare": false, 00:10:35.786 "compare_and_write": false, 00:10:35.786 "abort": true, 00:10:35.786 "seek_hole": false, 00:10:35.786 "seek_data": false, 00:10:35.786 "copy": true, 00:10:35.786 "nvme_iov_md": false 00:10:35.786 }, 00:10:35.786 "memory_domains": [ 00:10:35.786 { 00:10:35.786 "dma_device_id": "system", 00:10:35.786 "dma_device_type": 1 00:10:35.786 }, 00:10:35.786 { 00:10:35.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.786 "dma_device_type": 2 00:10:35.786 } 00:10:35.786 ], 00:10:35.786 "driver_specific": {} 00:10:35.786 } 00:10:35.786 ] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.786 [2024-09-30 14:09:40.229214] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:35.786 [2024-09-30 14:09:40.229263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:35.786 [2024-09-30 14:09:40.229282] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:35.786 [2024-09-30 14:09:40.231058] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:35.786 [2024-09-30 14:09:40.231107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.786 "name": "Existed_Raid", 00:10:35.786 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:35.786 "strip_size_kb": 0, 00:10:35.786 "state": "configuring", 00:10:35.786 "raid_level": "raid1", 00:10:35.786 "superblock": true, 00:10:35.786 "num_base_bdevs": 4, 00:10:35.786 "num_base_bdevs_discovered": 3, 00:10:35.786 "num_base_bdevs_operational": 4, 00:10:35.786 "base_bdevs_list": [ 00:10:35.786 { 00:10:35.786 "name": "BaseBdev1", 00:10:35.786 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.786 "is_configured": false, 00:10:35.786 "data_offset": 0, 00:10:35.786 "data_size": 0 00:10:35.786 }, 00:10:35.786 { 00:10:35.786 "name": "BaseBdev2", 00:10:35.786 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:35.786 "is_configured": true, 00:10:35.786 "data_offset": 2048, 00:10:35.786 "data_size": 63488 00:10:35.786 }, 00:10:35.786 { 00:10:35.786 "name": "BaseBdev3", 00:10:35.786 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:35.786 "is_configured": true, 00:10:35.786 "data_offset": 2048, 00:10:35.786 "data_size": 63488 00:10:35.786 }, 00:10:35.786 { 00:10:35.786 "name": "BaseBdev4", 00:10:35.786 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:35.786 "is_configured": true, 00:10:35.786 "data_offset": 2048, 00:10:35.786 "data_size": 63488 00:10:35.786 } 00:10:35.786 ] 00:10:35.786 }' 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.786 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.047 [2024-09-30 14:09:40.672461] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.047 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.307 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.307 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.307 "name": "Existed_Raid", 00:10:36.307 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:36.307 "strip_size_kb": 0, 00:10:36.307 "state": "configuring", 00:10:36.307 "raid_level": "raid1", 00:10:36.307 "superblock": true, 00:10:36.307 "num_base_bdevs": 4, 00:10:36.307 "num_base_bdevs_discovered": 2, 00:10:36.307 "num_base_bdevs_operational": 4, 00:10:36.307 "base_bdevs_list": [ 00:10:36.307 { 00:10:36.307 "name": "BaseBdev1", 00:10:36.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.307 "is_configured": false, 00:10:36.307 "data_offset": 0, 00:10:36.307 "data_size": 0 00:10:36.307 }, 00:10:36.307 { 00:10:36.307 "name": null, 00:10:36.307 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:36.307 "is_configured": false, 00:10:36.307 "data_offset": 0, 00:10:36.307 "data_size": 63488 00:10:36.307 }, 00:10:36.307 { 00:10:36.307 "name": "BaseBdev3", 00:10:36.307 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:36.307 "is_configured": true, 00:10:36.307 "data_offset": 2048, 00:10:36.307 "data_size": 63488 00:10:36.307 }, 00:10:36.307 { 00:10:36.307 "name": "BaseBdev4", 00:10:36.307 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:36.307 "is_configured": true, 00:10:36.307 "data_offset": 2048, 00:10:36.307 "data_size": 63488 00:10:36.307 } 00:10:36.307 ] 00:10:36.307 }' 00:10:36.307 14:09:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.307 14:09:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.568 [2024-09-30 14:09:41.198283] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:36.568 BaseBdev1 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.568 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.568 [ 00:10:36.568 { 00:10:36.568 "name": "BaseBdev1", 00:10:36.568 "aliases": [ 00:10:36.568 "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21" 00:10:36.568 ], 00:10:36.568 "product_name": "Malloc disk", 00:10:36.568 "block_size": 512, 00:10:36.568 "num_blocks": 65536, 00:10:36.568 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:36.568 "assigned_rate_limits": { 00:10:36.568 "rw_ios_per_sec": 0, 00:10:36.568 "rw_mbytes_per_sec": 0, 00:10:36.568 "r_mbytes_per_sec": 0, 00:10:36.568 "w_mbytes_per_sec": 0 00:10:36.568 }, 00:10:36.568 "claimed": true, 00:10:36.568 "claim_type": "exclusive_write", 00:10:36.568 "zoned": false, 00:10:36.568 "supported_io_types": { 00:10:36.828 "read": true, 00:10:36.828 "write": true, 00:10:36.828 "unmap": true, 00:10:36.828 "flush": true, 00:10:36.828 "reset": true, 00:10:36.828 "nvme_admin": false, 00:10:36.828 "nvme_io": false, 00:10:36.828 "nvme_io_md": false, 00:10:36.828 "write_zeroes": true, 00:10:36.828 "zcopy": true, 00:10:36.828 "get_zone_info": false, 00:10:36.828 "zone_management": false, 00:10:36.828 "zone_append": false, 00:10:36.828 "compare": false, 00:10:36.828 "compare_and_write": false, 00:10:36.828 "abort": true, 00:10:36.828 "seek_hole": false, 00:10:36.828 "seek_data": false, 00:10:36.828 "copy": true, 00:10:36.828 "nvme_iov_md": false 00:10:36.828 }, 00:10:36.828 "memory_domains": [ 00:10:36.828 { 00:10:36.828 "dma_device_id": "system", 00:10:36.828 "dma_device_type": 1 00:10:36.828 }, 00:10:36.828 { 00:10:36.828 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.828 "dma_device_type": 2 00:10:36.828 } 00:10:36.828 ], 00:10:36.829 "driver_specific": {} 00:10:36.829 } 00:10:36.829 ] 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.829 "name": "Existed_Raid", 00:10:36.829 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:36.829 "strip_size_kb": 0, 00:10:36.829 "state": "configuring", 00:10:36.829 "raid_level": "raid1", 00:10:36.829 "superblock": true, 00:10:36.829 "num_base_bdevs": 4, 00:10:36.829 "num_base_bdevs_discovered": 3, 00:10:36.829 "num_base_bdevs_operational": 4, 00:10:36.829 "base_bdevs_list": [ 00:10:36.829 { 00:10:36.829 "name": "BaseBdev1", 00:10:36.829 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:36.829 "is_configured": true, 00:10:36.829 "data_offset": 2048, 00:10:36.829 "data_size": 63488 00:10:36.829 }, 00:10:36.829 { 00:10:36.829 "name": null, 00:10:36.829 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:36.829 "is_configured": false, 00:10:36.829 "data_offset": 0, 00:10:36.829 "data_size": 63488 00:10:36.829 }, 00:10:36.829 { 00:10:36.829 "name": "BaseBdev3", 00:10:36.829 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:36.829 "is_configured": true, 00:10:36.829 "data_offset": 2048, 00:10:36.829 "data_size": 63488 00:10:36.829 }, 00:10:36.829 { 00:10:36.829 "name": "BaseBdev4", 00:10:36.829 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:36.829 "is_configured": true, 00:10:36.829 "data_offset": 2048, 00:10:36.829 "data_size": 63488 00:10:36.829 } 00:10:36.829 ] 00:10:36.829 }' 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.829 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.089 [2024-09-30 14:09:41.681489] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.089 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.090 "name": "Existed_Raid", 00:10:37.090 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:37.090 "strip_size_kb": 0, 00:10:37.090 "state": "configuring", 00:10:37.090 "raid_level": "raid1", 00:10:37.090 "superblock": true, 00:10:37.090 "num_base_bdevs": 4, 00:10:37.090 "num_base_bdevs_discovered": 2, 00:10:37.090 "num_base_bdevs_operational": 4, 00:10:37.090 "base_bdevs_list": [ 00:10:37.090 { 00:10:37.090 "name": "BaseBdev1", 00:10:37.090 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:37.090 "is_configured": true, 00:10:37.090 "data_offset": 2048, 00:10:37.090 "data_size": 63488 00:10:37.090 }, 00:10:37.090 { 00:10:37.090 "name": null, 00:10:37.090 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:37.090 "is_configured": false, 00:10:37.090 "data_offset": 0, 00:10:37.090 "data_size": 63488 00:10:37.090 }, 00:10:37.090 { 00:10:37.090 "name": null, 00:10:37.090 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:37.090 "is_configured": false, 00:10:37.090 "data_offset": 0, 00:10:37.090 "data_size": 63488 00:10:37.090 }, 00:10:37.090 { 00:10:37.090 "name": "BaseBdev4", 00:10:37.090 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:37.090 "is_configured": true, 00:10:37.090 "data_offset": 2048, 00:10:37.090 "data_size": 63488 00:10:37.090 } 00:10:37.090 ] 00:10:37.090 }' 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.090 14:09:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.660 [2024-09-30 14:09:42.200616] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.660 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.660 "name": "Existed_Raid", 00:10:37.660 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:37.660 "strip_size_kb": 0, 00:10:37.660 "state": "configuring", 00:10:37.660 "raid_level": "raid1", 00:10:37.660 "superblock": true, 00:10:37.661 "num_base_bdevs": 4, 00:10:37.661 "num_base_bdevs_discovered": 3, 00:10:37.661 "num_base_bdevs_operational": 4, 00:10:37.661 "base_bdevs_list": [ 00:10:37.661 { 00:10:37.661 "name": "BaseBdev1", 00:10:37.661 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:37.661 "is_configured": true, 00:10:37.661 "data_offset": 2048, 00:10:37.661 "data_size": 63488 00:10:37.661 }, 00:10:37.661 { 00:10:37.661 "name": null, 00:10:37.661 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:37.661 "is_configured": false, 00:10:37.661 "data_offset": 0, 00:10:37.661 "data_size": 63488 00:10:37.661 }, 00:10:37.661 { 00:10:37.661 "name": "BaseBdev3", 00:10:37.661 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:37.661 "is_configured": true, 00:10:37.661 "data_offset": 2048, 00:10:37.661 "data_size": 63488 00:10:37.661 }, 00:10:37.661 { 00:10:37.661 "name": "BaseBdev4", 00:10:37.661 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:37.661 "is_configured": true, 00:10:37.661 "data_offset": 2048, 00:10:37.661 "data_size": 63488 00:10:37.661 } 00:10:37.661 ] 00:10:37.661 }' 00:10:37.661 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.661 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.280 [2024-09-30 14:09:42.715771] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.280 "name": "Existed_Raid", 00:10:38.280 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:38.280 "strip_size_kb": 0, 00:10:38.280 "state": "configuring", 00:10:38.280 "raid_level": "raid1", 00:10:38.280 "superblock": true, 00:10:38.280 "num_base_bdevs": 4, 00:10:38.280 "num_base_bdevs_discovered": 2, 00:10:38.280 "num_base_bdevs_operational": 4, 00:10:38.280 "base_bdevs_list": [ 00:10:38.280 { 00:10:38.280 "name": null, 00:10:38.280 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:38.280 "is_configured": false, 00:10:38.280 "data_offset": 0, 00:10:38.280 "data_size": 63488 00:10:38.280 }, 00:10:38.280 { 00:10:38.280 "name": null, 00:10:38.280 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:38.280 "is_configured": false, 00:10:38.280 "data_offset": 0, 00:10:38.280 "data_size": 63488 00:10:38.280 }, 00:10:38.280 { 00:10:38.280 "name": "BaseBdev3", 00:10:38.280 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:38.280 "is_configured": true, 00:10:38.280 "data_offset": 2048, 00:10:38.280 "data_size": 63488 00:10:38.280 }, 00:10:38.280 { 00:10:38.280 "name": "BaseBdev4", 00:10:38.280 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:38.280 "is_configured": true, 00:10:38.280 "data_offset": 2048, 00:10:38.280 "data_size": 63488 00:10:38.280 } 00:10:38.280 ] 00:10:38.280 }' 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.280 14:09:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.540 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.800 [2024-09-30 14:09:43.201178] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.800 "name": "Existed_Raid", 00:10:38.800 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:38.800 "strip_size_kb": 0, 00:10:38.800 "state": "configuring", 00:10:38.800 "raid_level": "raid1", 00:10:38.800 "superblock": true, 00:10:38.800 "num_base_bdevs": 4, 00:10:38.800 "num_base_bdevs_discovered": 3, 00:10:38.800 "num_base_bdevs_operational": 4, 00:10:38.800 "base_bdevs_list": [ 00:10:38.800 { 00:10:38.800 "name": null, 00:10:38.800 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:38.800 "is_configured": false, 00:10:38.800 "data_offset": 0, 00:10:38.800 "data_size": 63488 00:10:38.800 }, 00:10:38.800 { 00:10:38.800 "name": "BaseBdev2", 00:10:38.800 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:38.800 "is_configured": true, 00:10:38.800 "data_offset": 2048, 00:10:38.800 "data_size": 63488 00:10:38.800 }, 00:10:38.800 { 00:10:38.800 "name": "BaseBdev3", 00:10:38.800 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:38.800 "is_configured": true, 00:10:38.800 "data_offset": 2048, 00:10:38.800 "data_size": 63488 00:10:38.800 }, 00:10:38.800 { 00:10:38.800 "name": "BaseBdev4", 00:10:38.800 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:38.800 "is_configured": true, 00:10:38.800 "data_offset": 2048, 00:10:38.800 "data_size": 63488 00:10:38.800 } 00:10:38.800 ] 00:10:38.800 }' 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.800 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.061 [2024-09-30 14:09:43.711029] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:39.061 [2024-09-30 14:09:43.711212] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:39.061 [2024-09-30 14:09:43.711231] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:39.061 NewBaseBdev 00:10:39.061 [2024-09-30 14:09:43.711501] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:10:39.061 [2024-09-30 14:09:43.711620] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:39.061 [2024-09-30 14:09:43.711639] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:39.061 [2024-09-30 14:09:43.711734] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.061 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.321 [ 00:10:39.321 { 00:10:39.321 "name": "NewBaseBdev", 00:10:39.321 "aliases": [ 00:10:39.321 "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21" 00:10:39.321 ], 00:10:39.321 "product_name": "Malloc disk", 00:10:39.321 "block_size": 512, 00:10:39.321 "num_blocks": 65536, 00:10:39.321 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:39.321 "assigned_rate_limits": { 00:10:39.321 "rw_ios_per_sec": 0, 00:10:39.321 "rw_mbytes_per_sec": 0, 00:10:39.321 "r_mbytes_per_sec": 0, 00:10:39.321 "w_mbytes_per_sec": 0 00:10:39.321 }, 00:10:39.321 "claimed": true, 00:10:39.321 "claim_type": "exclusive_write", 00:10:39.321 "zoned": false, 00:10:39.321 "supported_io_types": { 00:10:39.321 "read": true, 00:10:39.321 "write": true, 00:10:39.321 "unmap": true, 00:10:39.321 "flush": true, 00:10:39.321 "reset": true, 00:10:39.321 "nvme_admin": false, 00:10:39.321 "nvme_io": false, 00:10:39.321 "nvme_io_md": false, 00:10:39.321 "write_zeroes": true, 00:10:39.321 "zcopy": true, 00:10:39.321 "get_zone_info": false, 00:10:39.321 "zone_management": false, 00:10:39.321 "zone_append": false, 00:10:39.321 "compare": false, 00:10:39.321 "compare_and_write": false, 00:10:39.321 "abort": true, 00:10:39.321 "seek_hole": false, 00:10:39.321 "seek_data": false, 00:10:39.321 "copy": true, 00:10:39.321 "nvme_iov_md": false 00:10:39.321 }, 00:10:39.321 "memory_domains": [ 00:10:39.321 { 00:10:39.321 "dma_device_id": "system", 00:10:39.321 "dma_device_type": 1 00:10:39.321 }, 00:10:39.321 { 00:10:39.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.321 "dma_device_type": 2 00:10:39.321 } 00:10:39.321 ], 00:10:39.321 "driver_specific": {} 00:10:39.321 } 00:10:39.321 ] 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.321 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.321 "name": "Existed_Raid", 00:10:39.321 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:39.321 "strip_size_kb": 0, 00:10:39.321 "state": "online", 00:10:39.321 "raid_level": "raid1", 00:10:39.321 "superblock": true, 00:10:39.321 "num_base_bdevs": 4, 00:10:39.321 "num_base_bdevs_discovered": 4, 00:10:39.321 "num_base_bdevs_operational": 4, 00:10:39.321 "base_bdevs_list": [ 00:10:39.321 { 00:10:39.321 "name": "NewBaseBdev", 00:10:39.321 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:39.321 "is_configured": true, 00:10:39.321 "data_offset": 2048, 00:10:39.321 "data_size": 63488 00:10:39.321 }, 00:10:39.321 { 00:10:39.321 "name": "BaseBdev2", 00:10:39.321 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:39.321 "is_configured": true, 00:10:39.321 "data_offset": 2048, 00:10:39.321 "data_size": 63488 00:10:39.321 }, 00:10:39.321 { 00:10:39.321 "name": "BaseBdev3", 00:10:39.321 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:39.321 "is_configured": true, 00:10:39.321 "data_offset": 2048, 00:10:39.321 "data_size": 63488 00:10:39.321 }, 00:10:39.321 { 00:10:39.321 "name": "BaseBdev4", 00:10:39.321 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:39.321 "is_configured": true, 00:10:39.321 "data_offset": 2048, 00:10:39.321 "data_size": 63488 00:10:39.321 } 00:10:39.321 ] 00:10:39.321 }' 00:10:39.322 14:09:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.322 14:09:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:39.581 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.582 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.582 [2024-09-30 14:09:44.210516] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:39.582 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.842 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:39.842 "name": "Existed_Raid", 00:10:39.842 "aliases": [ 00:10:39.842 "287102dd-602a-41e4-9c0e-db377343dc91" 00:10:39.842 ], 00:10:39.842 "product_name": "Raid Volume", 00:10:39.842 "block_size": 512, 00:10:39.842 "num_blocks": 63488, 00:10:39.842 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:39.842 "assigned_rate_limits": { 00:10:39.842 "rw_ios_per_sec": 0, 00:10:39.842 "rw_mbytes_per_sec": 0, 00:10:39.842 "r_mbytes_per_sec": 0, 00:10:39.842 "w_mbytes_per_sec": 0 00:10:39.842 }, 00:10:39.842 "claimed": false, 00:10:39.842 "zoned": false, 00:10:39.842 "supported_io_types": { 00:10:39.842 "read": true, 00:10:39.842 "write": true, 00:10:39.842 "unmap": false, 00:10:39.842 "flush": false, 00:10:39.842 "reset": true, 00:10:39.842 "nvme_admin": false, 00:10:39.842 "nvme_io": false, 00:10:39.842 "nvme_io_md": false, 00:10:39.842 "write_zeroes": true, 00:10:39.842 "zcopy": false, 00:10:39.842 "get_zone_info": false, 00:10:39.842 "zone_management": false, 00:10:39.842 "zone_append": false, 00:10:39.842 "compare": false, 00:10:39.842 "compare_and_write": false, 00:10:39.842 "abort": false, 00:10:39.842 "seek_hole": false, 00:10:39.842 "seek_data": false, 00:10:39.842 "copy": false, 00:10:39.842 "nvme_iov_md": false 00:10:39.842 }, 00:10:39.842 "memory_domains": [ 00:10:39.842 { 00:10:39.842 "dma_device_id": "system", 00:10:39.842 "dma_device_type": 1 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.842 "dma_device_type": 2 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "system", 00:10:39.842 "dma_device_type": 1 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.842 "dma_device_type": 2 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "system", 00:10:39.842 "dma_device_type": 1 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.842 "dma_device_type": 2 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "system", 00:10:39.842 "dma_device_type": 1 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.842 "dma_device_type": 2 00:10:39.842 } 00:10:39.842 ], 00:10:39.842 "driver_specific": { 00:10:39.842 "raid": { 00:10:39.842 "uuid": "287102dd-602a-41e4-9c0e-db377343dc91", 00:10:39.842 "strip_size_kb": 0, 00:10:39.842 "state": "online", 00:10:39.842 "raid_level": "raid1", 00:10:39.842 "superblock": true, 00:10:39.842 "num_base_bdevs": 4, 00:10:39.842 "num_base_bdevs_discovered": 4, 00:10:39.842 "num_base_bdevs_operational": 4, 00:10:39.842 "base_bdevs_list": [ 00:10:39.842 { 00:10:39.842 "name": "NewBaseBdev", 00:10:39.842 "uuid": "e1a25cd8-8bc1-4ab5-a4a7-daee3e01aa21", 00:10:39.842 "is_configured": true, 00:10:39.842 "data_offset": 2048, 00:10:39.842 "data_size": 63488 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "name": "BaseBdev2", 00:10:39.842 "uuid": "3fded517-2e98-4ef5-8e32-97e40e08529d", 00:10:39.842 "is_configured": true, 00:10:39.842 "data_offset": 2048, 00:10:39.842 "data_size": 63488 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "name": "BaseBdev3", 00:10:39.842 "uuid": "aaaa5729-5484-4673-a639-c0151a44efdd", 00:10:39.842 "is_configured": true, 00:10:39.842 "data_offset": 2048, 00:10:39.842 "data_size": 63488 00:10:39.842 }, 00:10:39.842 { 00:10:39.842 "name": "BaseBdev4", 00:10:39.842 "uuid": "3af2c9dc-d0c6-4fdb-bd68-65349f398032", 00:10:39.843 "is_configured": true, 00:10:39.843 "data_offset": 2048, 00:10:39.843 "data_size": 63488 00:10:39.843 } 00:10:39.843 ] 00:10:39.843 } 00:10:39.843 } 00:10:39.843 }' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:39.843 BaseBdev2 00:10:39.843 BaseBdev3 00:10:39.843 BaseBdev4' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.843 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.103 [2024-09-30 14:09:44.521664] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:40.103 [2024-09-30 14:09:44.521696] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:40.103 [2024-09-30 14:09:44.521769] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:40.103 [2024-09-30 14:09:44.522011] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:40.103 [2024-09-30 14:09:44.522028] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 85653 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 85653 ']' 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 85653 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85653 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85653' 00:10:40.103 killing process with pid 85653 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 85653 00:10:40.103 [2024-09-30 14:09:44.571854] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:40.103 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 85653 00:10:40.103 [2024-09-30 14:09:44.611950] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:40.363 14:09:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:40.363 00:10:40.363 real 0m9.466s 00:10:40.363 user 0m16.094s 00:10:40.363 sys 0m2.098s 00:10:40.363 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:40.363 14:09:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.363 ************************************ 00:10:40.363 END TEST raid_state_function_test_sb 00:10:40.363 ************************************ 00:10:40.363 14:09:44 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:40.363 14:09:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:40.363 14:09:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:40.363 14:09:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:40.363 ************************************ 00:10:40.363 START TEST raid_superblock_test 00:10:40.363 ************************************ 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 4 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:40.363 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=86307 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 86307 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 86307 ']' 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:40.364 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:40.364 14:09:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.624 [2024-09-30 14:09:45.035099] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:40.624 [2024-09-30 14:09:45.035288] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86307 ] 00:10:40.624 [2024-09-30 14:09:45.172965] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:40.624 [2024-09-30 14:09:45.201320] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:40.624 [2024-09-30 14:09:45.247756] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:40.884 [2024-09-30 14:09:45.289891] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:40.884 [2024-09-30 14:09:45.289933] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.453 malloc1 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.453 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.453 [2024-09-30 14:09:45.887735] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:41.454 [2024-09-30 14:09:45.887822] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.454 [2024-09-30 14:09:45.887851] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:41.454 [2024-09-30 14:09:45.887869] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.454 [2024-09-30 14:09:45.889933] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.454 [2024-09-30 14:09:45.889967] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:41.454 pt1 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 malloc2 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 [2024-09-30 14:09:45.933117] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:41.454 [2024-09-30 14:09:45.933222] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.454 [2024-09-30 14:09:45.933265] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:41.454 [2024-09-30 14:09:45.933287] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.454 [2024-09-30 14:09:45.938069] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.454 [2024-09-30 14:09:45.938138] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:41.454 pt2 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 malloc3 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 [2024-09-30 14:09:45.963638] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:41.454 [2024-09-30 14:09:45.963684] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.454 [2024-09-30 14:09:45.963721] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:41.454 [2024-09-30 14:09:45.963731] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.454 [2024-09-30 14:09:45.965726] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.454 [2024-09-30 14:09:45.965758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:41.454 pt3 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 malloc4 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 [2024-09-30 14:09:45.991887] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:41.454 [2024-09-30 14:09:45.991934] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.454 [2024-09-30 14:09:45.991968] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:41.454 [2024-09-30 14:09:45.991977] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.454 [2024-09-30 14:09:45.993986] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.454 [2024-09-30 14:09:45.994017] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:41.454 pt4 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.454 14:09:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.454 [2024-09-30 14:09:46.003903] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:41.454 [2024-09-30 14:09:46.005640] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:41.454 [2024-09-30 14:09:46.005706] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:41.454 [2024-09-30 14:09:46.005747] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:41.454 [2024-09-30 14:09:46.005892] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:10:41.454 [2024-09-30 14:09:46.005910] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:41.454 [2024-09-30 14:09:46.006159] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:41.454 [2024-09-30 14:09:46.006313] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:10:41.454 [2024-09-30 14:09:46.006337] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:10:41.454 [2024-09-30 14:09:46.006450] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.454 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.455 "name": "raid_bdev1", 00:10:41.455 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:41.455 "strip_size_kb": 0, 00:10:41.455 "state": "online", 00:10:41.455 "raid_level": "raid1", 00:10:41.455 "superblock": true, 00:10:41.455 "num_base_bdevs": 4, 00:10:41.455 "num_base_bdevs_discovered": 4, 00:10:41.455 "num_base_bdevs_operational": 4, 00:10:41.455 "base_bdevs_list": [ 00:10:41.455 { 00:10:41.455 "name": "pt1", 00:10:41.455 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:41.455 "is_configured": true, 00:10:41.455 "data_offset": 2048, 00:10:41.455 "data_size": 63488 00:10:41.455 }, 00:10:41.455 { 00:10:41.455 "name": "pt2", 00:10:41.455 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.455 "is_configured": true, 00:10:41.455 "data_offset": 2048, 00:10:41.455 "data_size": 63488 00:10:41.455 }, 00:10:41.455 { 00:10:41.455 "name": "pt3", 00:10:41.455 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.455 "is_configured": true, 00:10:41.455 "data_offset": 2048, 00:10:41.455 "data_size": 63488 00:10:41.455 }, 00:10:41.455 { 00:10:41.455 "name": "pt4", 00:10:41.455 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:41.455 "is_configured": true, 00:10:41.455 "data_offset": 2048, 00:10:41.455 "data_size": 63488 00:10:41.455 } 00:10:41.455 ] 00:10:41.455 }' 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.455 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.025 [2024-09-30 14:09:46.443406] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:42.025 "name": "raid_bdev1", 00:10:42.025 "aliases": [ 00:10:42.025 "12da7736-8e54-4fa4-b0a1-dda05adc1df9" 00:10:42.025 ], 00:10:42.025 "product_name": "Raid Volume", 00:10:42.025 "block_size": 512, 00:10:42.025 "num_blocks": 63488, 00:10:42.025 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:42.025 "assigned_rate_limits": { 00:10:42.025 "rw_ios_per_sec": 0, 00:10:42.025 "rw_mbytes_per_sec": 0, 00:10:42.025 "r_mbytes_per_sec": 0, 00:10:42.025 "w_mbytes_per_sec": 0 00:10:42.025 }, 00:10:42.025 "claimed": false, 00:10:42.025 "zoned": false, 00:10:42.025 "supported_io_types": { 00:10:42.025 "read": true, 00:10:42.025 "write": true, 00:10:42.025 "unmap": false, 00:10:42.025 "flush": false, 00:10:42.025 "reset": true, 00:10:42.025 "nvme_admin": false, 00:10:42.025 "nvme_io": false, 00:10:42.025 "nvme_io_md": false, 00:10:42.025 "write_zeroes": true, 00:10:42.025 "zcopy": false, 00:10:42.025 "get_zone_info": false, 00:10:42.025 "zone_management": false, 00:10:42.025 "zone_append": false, 00:10:42.025 "compare": false, 00:10:42.025 "compare_and_write": false, 00:10:42.025 "abort": false, 00:10:42.025 "seek_hole": false, 00:10:42.025 "seek_data": false, 00:10:42.025 "copy": false, 00:10:42.025 "nvme_iov_md": false 00:10:42.025 }, 00:10:42.025 "memory_domains": [ 00:10:42.025 { 00:10:42.025 "dma_device_id": "system", 00:10:42.025 "dma_device_type": 1 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.025 "dma_device_type": 2 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "system", 00:10:42.025 "dma_device_type": 1 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.025 "dma_device_type": 2 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "system", 00:10:42.025 "dma_device_type": 1 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.025 "dma_device_type": 2 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "system", 00:10:42.025 "dma_device_type": 1 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.025 "dma_device_type": 2 00:10:42.025 } 00:10:42.025 ], 00:10:42.025 "driver_specific": { 00:10:42.025 "raid": { 00:10:42.025 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:42.025 "strip_size_kb": 0, 00:10:42.025 "state": "online", 00:10:42.025 "raid_level": "raid1", 00:10:42.025 "superblock": true, 00:10:42.025 "num_base_bdevs": 4, 00:10:42.025 "num_base_bdevs_discovered": 4, 00:10:42.025 "num_base_bdevs_operational": 4, 00:10:42.025 "base_bdevs_list": [ 00:10:42.025 { 00:10:42.025 "name": "pt1", 00:10:42.025 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:42.025 "is_configured": true, 00:10:42.025 "data_offset": 2048, 00:10:42.025 "data_size": 63488 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "name": "pt2", 00:10:42.025 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.025 "is_configured": true, 00:10:42.025 "data_offset": 2048, 00:10:42.025 "data_size": 63488 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "name": "pt3", 00:10:42.025 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.025 "is_configured": true, 00:10:42.025 "data_offset": 2048, 00:10:42.025 "data_size": 63488 00:10:42.025 }, 00:10:42.025 { 00:10:42.025 "name": "pt4", 00:10:42.025 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:42.025 "is_configured": true, 00:10:42.025 "data_offset": 2048, 00:10:42.025 "data_size": 63488 00:10:42.025 } 00:10:42.025 ] 00:10:42.025 } 00:10:42.025 } 00:10:42.025 }' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:42.025 pt2 00:10:42.025 pt3 00:10:42.025 pt4' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.025 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.026 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.286 [2024-09-30 14:09:46.690946] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=12da7736-8e54-4fa4-b0a1-dda05adc1df9 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 12da7736-8e54-4fa4-b0a1-dda05adc1df9 ']' 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.286 [2024-09-30 14:09:46.738580] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:42.286 [2024-09-30 14:09:46.738605] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:42.286 [2024-09-30 14:09:46.738678] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:42.286 [2024-09-30 14:09:46.738761] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:42.286 [2024-09-30 14:09:46.738774] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.286 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 [2024-09-30 14:09:46.898327] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:42.287 [2024-09-30 14:09:46.900107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:42.287 [2024-09-30 14:09:46.900155] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:42.287 [2024-09-30 14:09:46.900183] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:42.287 [2024-09-30 14:09:46.900223] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:42.287 [2024-09-30 14:09:46.900264] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:42.287 [2024-09-30 14:09:46.900302] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:42.287 [2024-09-30 14:09:46.900319] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:42.287 [2024-09-30 14:09:46.900332] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:42.287 [2024-09-30 14:09:46.900350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:10:42.287 request: 00:10:42.287 { 00:10:42.287 "name": "raid_bdev1", 00:10:42.287 "raid_level": "raid1", 00:10:42.287 "base_bdevs": [ 00:10:42.287 "malloc1", 00:10:42.287 "malloc2", 00:10:42.287 "malloc3", 00:10:42.287 "malloc4" 00:10:42.287 ], 00:10:42.287 "superblock": false, 00:10:42.287 "method": "bdev_raid_create", 00:10:42.287 "req_id": 1 00:10:42.287 } 00:10:42.287 Got JSON-RPC error response 00:10:42.287 response: 00:10:42.287 { 00:10:42.287 "code": -17, 00:10:42.287 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:42.287 } 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.287 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.547 [2024-09-30 14:09:46.942233] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:42.547 [2024-09-30 14:09:46.942285] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.547 [2024-09-30 14:09:46.942302] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:42.547 [2024-09-30 14:09:46.942312] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.547 [2024-09-30 14:09:46.944382] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.547 [2024-09-30 14:09:46.944419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:42.547 [2024-09-30 14:09:46.944477] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:42.547 [2024-09-30 14:09:46.944532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:42.547 pt1 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.547 "name": "raid_bdev1", 00:10:42.547 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:42.547 "strip_size_kb": 0, 00:10:42.547 "state": "configuring", 00:10:42.547 "raid_level": "raid1", 00:10:42.547 "superblock": true, 00:10:42.547 "num_base_bdevs": 4, 00:10:42.547 "num_base_bdevs_discovered": 1, 00:10:42.547 "num_base_bdevs_operational": 4, 00:10:42.547 "base_bdevs_list": [ 00:10:42.547 { 00:10:42.547 "name": "pt1", 00:10:42.547 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:42.547 "is_configured": true, 00:10:42.547 "data_offset": 2048, 00:10:42.547 "data_size": 63488 00:10:42.547 }, 00:10:42.547 { 00:10:42.547 "name": null, 00:10:42.547 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.547 "is_configured": false, 00:10:42.547 "data_offset": 2048, 00:10:42.547 "data_size": 63488 00:10:42.547 }, 00:10:42.547 { 00:10:42.547 "name": null, 00:10:42.547 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.547 "is_configured": false, 00:10:42.547 "data_offset": 2048, 00:10:42.547 "data_size": 63488 00:10:42.547 }, 00:10:42.547 { 00:10:42.547 "name": null, 00:10:42.547 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:42.547 "is_configured": false, 00:10:42.547 "data_offset": 2048, 00:10:42.547 "data_size": 63488 00:10:42.547 } 00:10:42.547 ] 00:10:42.547 }' 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.547 14:09:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.807 [2024-09-30 14:09:47.409438] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:42.807 [2024-09-30 14:09:47.409519] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.807 [2024-09-30 14:09:47.409543] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:42.807 [2024-09-30 14:09:47.409553] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.807 [2024-09-30 14:09:47.409908] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.807 [2024-09-30 14:09:47.409933] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:42.807 [2024-09-30 14:09:47.409996] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:42.807 [2024-09-30 14:09:47.410032] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:42.807 pt2 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.807 [2024-09-30 14:09:47.417439] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.807 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.808 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.067 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.067 "name": "raid_bdev1", 00:10:43.067 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:43.067 "strip_size_kb": 0, 00:10:43.067 "state": "configuring", 00:10:43.067 "raid_level": "raid1", 00:10:43.067 "superblock": true, 00:10:43.067 "num_base_bdevs": 4, 00:10:43.067 "num_base_bdevs_discovered": 1, 00:10:43.067 "num_base_bdevs_operational": 4, 00:10:43.067 "base_bdevs_list": [ 00:10:43.067 { 00:10:43.067 "name": "pt1", 00:10:43.067 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:43.067 "is_configured": true, 00:10:43.067 "data_offset": 2048, 00:10:43.067 "data_size": 63488 00:10:43.067 }, 00:10:43.067 { 00:10:43.067 "name": null, 00:10:43.067 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.067 "is_configured": false, 00:10:43.067 "data_offset": 0, 00:10:43.067 "data_size": 63488 00:10:43.067 }, 00:10:43.067 { 00:10:43.067 "name": null, 00:10:43.067 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.067 "is_configured": false, 00:10:43.067 "data_offset": 2048, 00:10:43.067 "data_size": 63488 00:10:43.067 }, 00:10:43.067 { 00:10:43.068 "name": null, 00:10:43.068 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:43.068 "is_configured": false, 00:10:43.068 "data_offset": 2048, 00:10:43.068 "data_size": 63488 00:10:43.068 } 00:10:43.068 ] 00:10:43.068 }' 00:10:43.068 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.068 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.328 [2024-09-30 14:09:47.880668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:43.328 [2024-09-30 14:09:47.880729] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:43.328 [2024-09-30 14:09:47.880747] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:43.328 [2024-09-30 14:09:47.880757] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:43.328 [2024-09-30 14:09:47.881146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:43.328 [2024-09-30 14:09:47.881170] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:43.328 [2024-09-30 14:09:47.881244] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:43.328 [2024-09-30 14:09:47.881280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:43.328 pt2 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.328 [2024-09-30 14:09:47.892599] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:43.328 [2024-09-30 14:09:47.892651] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:43.328 [2024-09-30 14:09:47.892668] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:43.328 [2024-09-30 14:09:47.892677] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:43.328 [2024-09-30 14:09:47.892996] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:43.328 [2024-09-30 14:09:47.893019] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:43.328 [2024-09-30 14:09:47.893076] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:43.328 [2024-09-30 14:09:47.893104] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:43.328 pt3 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.328 [2024-09-30 14:09:47.904570] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:43.328 [2024-09-30 14:09:47.904614] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:43.328 [2024-09-30 14:09:47.904628] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:43.328 [2024-09-30 14:09:47.904637] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:43.328 [2024-09-30 14:09:47.904926] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:43.328 [2024-09-30 14:09:47.904948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:43.328 [2024-09-30 14:09:47.904999] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:43.328 [2024-09-30 14:09:47.905022] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:43.328 [2024-09-30 14:09:47.905124] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:43.328 [2024-09-30 14:09:47.905137] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:43.328 [2024-09-30 14:09:47.905365] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:10:43.328 [2024-09-30 14:09:47.905496] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:43.328 [2024-09-30 14:09:47.905515] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:10:43.328 [2024-09-30 14:09:47.905610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:43.328 pt4 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.328 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.328 "name": "raid_bdev1", 00:10:43.328 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:43.328 "strip_size_kb": 0, 00:10:43.328 "state": "online", 00:10:43.329 "raid_level": "raid1", 00:10:43.329 "superblock": true, 00:10:43.329 "num_base_bdevs": 4, 00:10:43.329 "num_base_bdevs_discovered": 4, 00:10:43.329 "num_base_bdevs_operational": 4, 00:10:43.329 "base_bdevs_list": [ 00:10:43.329 { 00:10:43.329 "name": "pt1", 00:10:43.329 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:43.329 "is_configured": true, 00:10:43.329 "data_offset": 2048, 00:10:43.329 "data_size": 63488 00:10:43.329 }, 00:10:43.329 { 00:10:43.329 "name": "pt2", 00:10:43.329 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.329 "is_configured": true, 00:10:43.329 "data_offset": 2048, 00:10:43.329 "data_size": 63488 00:10:43.329 }, 00:10:43.329 { 00:10:43.329 "name": "pt3", 00:10:43.329 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.329 "is_configured": true, 00:10:43.329 "data_offset": 2048, 00:10:43.329 "data_size": 63488 00:10:43.329 }, 00:10:43.329 { 00:10:43.329 "name": "pt4", 00:10:43.329 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:43.329 "is_configured": true, 00:10:43.329 "data_offset": 2048, 00:10:43.329 "data_size": 63488 00:10:43.329 } 00:10:43.329 ] 00:10:43.329 }' 00:10:43.329 14:09:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.329 14:09:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.898 [2024-09-30 14:09:48.380084] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:43.898 "name": "raid_bdev1", 00:10:43.898 "aliases": [ 00:10:43.898 "12da7736-8e54-4fa4-b0a1-dda05adc1df9" 00:10:43.898 ], 00:10:43.898 "product_name": "Raid Volume", 00:10:43.898 "block_size": 512, 00:10:43.898 "num_blocks": 63488, 00:10:43.898 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:43.898 "assigned_rate_limits": { 00:10:43.898 "rw_ios_per_sec": 0, 00:10:43.898 "rw_mbytes_per_sec": 0, 00:10:43.898 "r_mbytes_per_sec": 0, 00:10:43.898 "w_mbytes_per_sec": 0 00:10:43.898 }, 00:10:43.898 "claimed": false, 00:10:43.898 "zoned": false, 00:10:43.898 "supported_io_types": { 00:10:43.898 "read": true, 00:10:43.898 "write": true, 00:10:43.898 "unmap": false, 00:10:43.898 "flush": false, 00:10:43.898 "reset": true, 00:10:43.898 "nvme_admin": false, 00:10:43.898 "nvme_io": false, 00:10:43.898 "nvme_io_md": false, 00:10:43.898 "write_zeroes": true, 00:10:43.898 "zcopy": false, 00:10:43.898 "get_zone_info": false, 00:10:43.898 "zone_management": false, 00:10:43.898 "zone_append": false, 00:10:43.898 "compare": false, 00:10:43.898 "compare_and_write": false, 00:10:43.898 "abort": false, 00:10:43.898 "seek_hole": false, 00:10:43.898 "seek_data": false, 00:10:43.898 "copy": false, 00:10:43.898 "nvme_iov_md": false 00:10:43.898 }, 00:10:43.898 "memory_domains": [ 00:10:43.898 { 00:10:43.898 "dma_device_id": "system", 00:10:43.898 "dma_device_type": 1 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.898 "dma_device_type": 2 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "system", 00:10:43.898 "dma_device_type": 1 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.898 "dma_device_type": 2 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "system", 00:10:43.898 "dma_device_type": 1 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.898 "dma_device_type": 2 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "system", 00:10:43.898 "dma_device_type": 1 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.898 "dma_device_type": 2 00:10:43.898 } 00:10:43.898 ], 00:10:43.898 "driver_specific": { 00:10:43.898 "raid": { 00:10:43.898 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:43.898 "strip_size_kb": 0, 00:10:43.898 "state": "online", 00:10:43.898 "raid_level": "raid1", 00:10:43.898 "superblock": true, 00:10:43.898 "num_base_bdevs": 4, 00:10:43.898 "num_base_bdevs_discovered": 4, 00:10:43.898 "num_base_bdevs_operational": 4, 00:10:43.898 "base_bdevs_list": [ 00:10:43.898 { 00:10:43.898 "name": "pt1", 00:10:43.898 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:43.898 "is_configured": true, 00:10:43.898 "data_offset": 2048, 00:10:43.898 "data_size": 63488 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "name": "pt2", 00:10:43.898 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.898 "is_configured": true, 00:10:43.898 "data_offset": 2048, 00:10:43.898 "data_size": 63488 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "name": "pt3", 00:10:43.898 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.898 "is_configured": true, 00:10:43.898 "data_offset": 2048, 00:10:43.898 "data_size": 63488 00:10:43.898 }, 00:10:43.898 { 00:10:43.898 "name": "pt4", 00:10:43.898 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:43.898 "is_configured": true, 00:10:43.898 "data_offset": 2048, 00:10:43.898 "data_size": 63488 00:10:43.898 } 00:10:43.898 ] 00:10:43.898 } 00:10:43.898 } 00:10:43.898 }' 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:43.898 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:43.898 pt2 00:10:43.898 pt3 00:10:43.898 pt4' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.899 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 [2024-09-30 14:09:48.687519] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 12da7736-8e54-4fa4-b0a1-dda05adc1df9 '!=' 12da7736-8e54-4fa4-b0a1-dda05adc1df9 ']' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 [2024-09-30 14:09:48.735225] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.159 "name": "raid_bdev1", 00:10:44.159 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:44.159 "strip_size_kb": 0, 00:10:44.159 "state": "online", 00:10:44.159 "raid_level": "raid1", 00:10:44.159 "superblock": true, 00:10:44.159 "num_base_bdevs": 4, 00:10:44.159 "num_base_bdevs_discovered": 3, 00:10:44.159 "num_base_bdevs_operational": 3, 00:10:44.159 "base_bdevs_list": [ 00:10:44.159 { 00:10:44.159 "name": null, 00:10:44.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.159 "is_configured": false, 00:10:44.159 "data_offset": 0, 00:10:44.159 "data_size": 63488 00:10:44.159 }, 00:10:44.159 { 00:10:44.159 "name": "pt2", 00:10:44.159 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:44.159 "is_configured": true, 00:10:44.159 "data_offset": 2048, 00:10:44.159 "data_size": 63488 00:10:44.159 }, 00:10:44.159 { 00:10:44.159 "name": "pt3", 00:10:44.159 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:44.159 "is_configured": true, 00:10:44.159 "data_offset": 2048, 00:10:44.159 "data_size": 63488 00:10:44.159 }, 00:10:44.159 { 00:10:44.159 "name": "pt4", 00:10:44.159 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:44.159 "is_configured": true, 00:10:44.159 "data_offset": 2048, 00:10:44.159 "data_size": 63488 00:10:44.159 } 00:10:44.159 ] 00:10:44.159 }' 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.159 14:09:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 [2024-09-30 14:09:49.162466] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:44.730 [2024-09-30 14:09:49.162509] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:44.730 [2024-09-30 14:09:49.162585] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:44.730 [2024-09-30 14:09:49.162660] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:44.730 [2024-09-30 14:09:49.162675] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 [2024-09-30 14:09:49.246314] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:44.730 [2024-09-30 14:09:49.246366] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.730 [2024-09-30 14:09:49.246384] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:44.730 [2024-09-30 14:09:49.246392] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.730 [2024-09-30 14:09:49.248555] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.730 [2024-09-30 14:09:49.248590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:44.730 [2024-09-30 14:09:49.248657] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:44.730 [2024-09-30 14:09:49.248696] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:44.730 pt2 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.730 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.730 "name": "raid_bdev1", 00:10:44.730 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:44.730 "strip_size_kb": 0, 00:10:44.730 "state": "configuring", 00:10:44.730 "raid_level": "raid1", 00:10:44.730 "superblock": true, 00:10:44.730 "num_base_bdevs": 4, 00:10:44.730 "num_base_bdevs_discovered": 1, 00:10:44.730 "num_base_bdevs_operational": 3, 00:10:44.730 "base_bdevs_list": [ 00:10:44.730 { 00:10:44.730 "name": null, 00:10:44.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.730 "is_configured": false, 00:10:44.730 "data_offset": 2048, 00:10:44.730 "data_size": 63488 00:10:44.730 }, 00:10:44.730 { 00:10:44.730 "name": "pt2", 00:10:44.730 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:44.730 "is_configured": true, 00:10:44.730 "data_offset": 2048, 00:10:44.730 "data_size": 63488 00:10:44.730 }, 00:10:44.730 { 00:10:44.730 "name": null, 00:10:44.730 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:44.730 "is_configured": false, 00:10:44.730 "data_offset": 2048, 00:10:44.730 "data_size": 63488 00:10:44.730 }, 00:10:44.731 { 00:10:44.731 "name": null, 00:10:44.731 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:44.731 "is_configured": false, 00:10:44.731 "data_offset": 2048, 00:10:44.731 "data_size": 63488 00:10:44.731 } 00:10:44.731 ] 00:10:44.731 }' 00:10:44.731 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.731 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.300 [2024-09-30 14:09:49.653631] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:45.300 [2024-09-30 14:09:49.653678] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:45.300 [2024-09-30 14:09:49.653699] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:10:45.300 [2024-09-30 14:09:49.653708] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:45.300 [2024-09-30 14:09:49.654080] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:45.300 [2024-09-30 14:09:49.654103] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:45.300 [2024-09-30 14:09:49.654168] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:45.300 [2024-09-30 14:09:49.654206] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:45.300 pt3 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.300 "name": "raid_bdev1", 00:10:45.300 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:45.300 "strip_size_kb": 0, 00:10:45.300 "state": "configuring", 00:10:45.300 "raid_level": "raid1", 00:10:45.300 "superblock": true, 00:10:45.300 "num_base_bdevs": 4, 00:10:45.300 "num_base_bdevs_discovered": 2, 00:10:45.300 "num_base_bdevs_operational": 3, 00:10:45.300 "base_bdevs_list": [ 00:10:45.300 { 00:10:45.300 "name": null, 00:10:45.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.300 "is_configured": false, 00:10:45.300 "data_offset": 2048, 00:10:45.300 "data_size": 63488 00:10:45.300 }, 00:10:45.300 { 00:10:45.300 "name": "pt2", 00:10:45.300 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:45.300 "is_configured": true, 00:10:45.300 "data_offset": 2048, 00:10:45.300 "data_size": 63488 00:10:45.300 }, 00:10:45.300 { 00:10:45.300 "name": "pt3", 00:10:45.300 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:45.300 "is_configured": true, 00:10:45.300 "data_offset": 2048, 00:10:45.300 "data_size": 63488 00:10:45.300 }, 00:10:45.300 { 00:10:45.300 "name": null, 00:10:45.300 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:45.300 "is_configured": false, 00:10:45.300 "data_offset": 2048, 00:10:45.300 "data_size": 63488 00:10:45.300 } 00:10:45.300 ] 00:10:45.300 }' 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.300 14:09:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.560 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.561 [2024-09-30 14:09:50.100890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:45.561 [2024-09-30 14:09:50.100967] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:45.561 [2024-09-30 14:09:50.100989] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:10:45.561 [2024-09-30 14:09:50.100998] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:45.561 [2024-09-30 14:09:50.101377] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:45.561 [2024-09-30 14:09:50.101401] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:45.561 [2024-09-30 14:09:50.101477] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:45.561 [2024-09-30 14:09:50.101526] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:45.561 [2024-09-30 14:09:50.101638] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:45.561 [2024-09-30 14:09:50.101654] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:45.561 [2024-09-30 14:09:50.101888] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:10:45.561 [2024-09-30 14:09:50.102038] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:45.561 [2024-09-30 14:09:50.102058] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:45.561 [2024-09-30 14:09:50.102160] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:45.561 pt4 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.561 "name": "raid_bdev1", 00:10:45.561 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:45.561 "strip_size_kb": 0, 00:10:45.561 "state": "online", 00:10:45.561 "raid_level": "raid1", 00:10:45.561 "superblock": true, 00:10:45.561 "num_base_bdevs": 4, 00:10:45.561 "num_base_bdevs_discovered": 3, 00:10:45.561 "num_base_bdevs_operational": 3, 00:10:45.561 "base_bdevs_list": [ 00:10:45.561 { 00:10:45.561 "name": null, 00:10:45.561 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.561 "is_configured": false, 00:10:45.561 "data_offset": 2048, 00:10:45.561 "data_size": 63488 00:10:45.561 }, 00:10:45.561 { 00:10:45.561 "name": "pt2", 00:10:45.561 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:45.561 "is_configured": true, 00:10:45.561 "data_offset": 2048, 00:10:45.561 "data_size": 63488 00:10:45.561 }, 00:10:45.561 { 00:10:45.561 "name": "pt3", 00:10:45.561 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:45.561 "is_configured": true, 00:10:45.561 "data_offset": 2048, 00:10:45.561 "data_size": 63488 00:10:45.561 }, 00:10:45.561 { 00:10:45.561 "name": "pt4", 00:10:45.561 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:45.561 "is_configured": true, 00:10:45.561 "data_offset": 2048, 00:10:45.561 "data_size": 63488 00:10:45.561 } 00:10:45.561 ] 00:10:45.561 }' 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.561 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.821 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:45.821 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.821 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.821 [2024-09-30 14:09:50.476221] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:45.821 [2024-09-30 14:09:50.476256] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:45.821 [2024-09-30 14:09:50.476339] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:45.821 [2024-09-30 14:09:50.476418] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:45.821 [2024-09-30 14:09:50.476436] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.081 [2024-09-30 14:09:50.548127] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:46.081 [2024-09-30 14:09:50.548187] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.081 [2024-09-30 14:09:50.548209] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:10:46.081 [2024-09-30 14:09:50.548220] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.081 [2024-09-30 14:09:50.550279] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.081 [2024-09-30 14:09:50.550318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:46.081 [2024-09-30 14:09:50.550384] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:46.081 [2024-09-30 14:09:50.550426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:46.081 [2024-09-30 14:09:50.550546] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:46.081 [2024-09-30 14:09:50.550576] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:46.081 [2024-09-30 14:09:50.550598] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:10:46.081 [2024-09-30 14:09:50.550642] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:46.081 [2024-09-30 14:09:50.550723] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:46.081 pt1 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.081 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.082 "name": "raid_bdev1", 00:10:46.082 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:46.082 "strip_size_kb": 0, 00:10:46.082 "state": "configuring", 00:10:46.082 "raid_level": "raid1", 00:10:46.082 "superblock": true, 00:10:46.082 "num_base_bdevs": 4, 00:10:46.082 "num_base_bdevs_discovered": 2, 00:10:46.082 "num_base_bdevs_operational": 3, 00:10:46.082 "base_bdevs_list": [ 00:10:46.082 { 00:10:46.082 "name": null, 00:10:46.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.082 "is_configured": false, 00:10:46.082 "data_offset": 2048, 00:10:46.082 "data_size": 63488 00:10:46.082 }, 00:10:46.082 { 00:10:46.082 "name": "pt2", 00:10:46.082 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:46.082 "is_configured": true, 00:10:46.082 "data_offset": 2048, 00:10:46.082 "data_size": 63488 00:10:46.082 }, 00:10:46.082 { 00:10:46.082 "name": "pt3", 00:10:46.082 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:46.082 "is_configured": true, 00:10:46.082 "data_offset": 2048, 00:10:46.082 "data_size": 63488 00:10:46.082 }, 00:10:46.082 { 00:10:46.082 "name": null, 00:10:46.082 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:46.082 "is_configured": false, 00:10:46.082 "data_offset": 2048, 00:10:46.082 "data_size": 63488 00:10:46.082 } 00:10:46.082 ] 00:10:46.082 }' 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.082 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.341 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:46.341 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.341 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.341 14:09:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:46.341 14:09:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.602 [2024-09-30 14:09:51.019313] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:46.602 [2024-09-30 14:09:51.019367] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.602 [2024-09-30 14:09:51.019403] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:10:46.602 [2024-09-30 14:09:51.019412] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.602 [2024-09-30 14:09:51.019788] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.602 [2024-09-30 14:09:51.019813] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:46.602 [2024-09-30 14:09:51.019881] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:46.602 [2024-09-30 14:09:51.019903] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:46.602 [2024-09-30 14:09:51.019999] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:10:46.602 [2024-09-30 14:09:51.020013] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:46.602 [2024-09-30 14:09:51.020256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:46.602 [2024-09-30 14:09:51.020373] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:10:46.602 [2024-09-30 14:09:51.020389] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:10:46.602 [2024-09-30 14:09:51.020504] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.602 pt4 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.602 "name": "raid_bdev1", 00:10:46.602 "uuid": "12da7736-8e54-4fa4-b0a1-dda05adc1df9", 00:10:46.602 "strip_size_kb": 0, 00:10:46.602 "state": "online", 00:10:46.602 "raid_level": "raid1", 00:10:46.602 "superblock": true, 00:10:46.602 "num_base_bdevs": 4, 00:10:46.602 "num_base_bdevs_discovered": 3, 00:10:46.602 "num_base_bdevs_operational": 3, 00:10:46.602 "base_bdevs_list": [ 00:10:46.602 { 00:10:46.602 "name": null, 00:10:46.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.602 "is_configured": false, 00:10:46.602 "data_offset": 2048, 00:10:46.602 "data_size": 63488 00:10:46.602 }, 00:10:46.602 { 00:10:46.602 "name": "pt2", 00:10:46.602 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:46.602 "is_configured": true, 00:10:46.602 "data_offset": 2048, 00:10:46.602 "data_size": 63488 00:10:46.602 }, 00:10:46.602 { 00:10:46.602 "name": "pt3", 00:10:46.602 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:46.602 "is_configured": true, 00:10:46.602 "data_offset": 2048, 00:10:46.602 "data_size": 63488 00:10:46.602 }, 00:10:46.602 { 00:10:46.602 "name": "pt4", 00:10:46.602 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:46.602 "is_configured": true, 00:10:46.602 "data_offset": 2048, 00:10:46.602 "data_size": 63488 00:10:46.602 } 00:10:46.602 ] 00:10:46.602 }' 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.602 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:46.862 [2024-09-30 14:09:51.482781] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:46.862 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 12da7736-8e54-4fa4-b0a1-dda05adc1df9 '!=' 12da7736-8e54-4fa4-b0a1-dda05adc1df9 ']' 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 86307 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 86307 ']' 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 86307 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86307 00:10:47.122 killing process with pid 86307 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86307' 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 86307 00:10:47.122 [2024-09-30 14:09:51.555856] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:47.122 [2024-09-30 14:09:51.555948] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:47.122 [2024-09-30 14:09:51.556022] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:47.122 [2024-09-30 14:09:51.556034] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:10:47.122 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 86307 00:10:47.122 [2024-09-30 14:09:51.597104] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:47.382 14:09:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:47.382 00:10:47.382 real 0m6.911s 00:10:47.382 user 0m11.585s 00:10:47.382 sys 0m1.544s 00:10:47.382 ************************************ 00:10:47.382 END TEST raid_superblock_test 00:10:47.382 ************************************ 00:10:47.382 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:47.382 14:09:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.382 14:09:51 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:10:47.382 14:09:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:47.382 14:09:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:47.382 14:09:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:47.382 ************************************ 00:10:47.382 START TEST raid_read_error_test 00:10:47.382 ************************************ 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 read 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:47.382 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.T1Fb57H9Pk 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=86777 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 86777 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 86777 ']' 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:47.383 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:47.383 14:09:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.643 [2024-09-30 14:09:52.041919] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:47.643 [2024-09-30 14:09:52.042136] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86777 ] 00:10:47.643 [2024-09-30 14:09:52.178098] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:47.643 [2024-09-30 14:09:52.207871] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:47.643 [2024-09-30 14:09:52.252590] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:47.643 [2024-09-30 14:09:52.294163] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:47.643 [2024-09-30 14:09:52.294281] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.212 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 BaseBdev1_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 true 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 [2024-09-30 14:09:52.900013] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:48.473 [2024-09-30 14:09:52.900075] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.473 [2024-09-30 14:09:52.900099] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:48.473 [2024-09-30 14:09:52.900112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.473 [2024-09-30 14:09:52.902172] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.473 [2024-09-30 14:09:52.902209] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:48.473 BaseBdev1 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 BaseBdev2_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 true 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 [2024-09-30 14:09:52.951282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:48.473 [2024-09-30 14:09:52.951396] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.473 [2024-09-30 14:09:52.951416] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:48.473 [2024-09-30 14:09:52.951426] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.473 [2024-09-30 14:09:52.953449] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.473 [2024-09-30 14:09:52.953493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:48.473 BaseBdev2 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 BaseBdev3_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 true 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 [2024-09-30 14:09:52.991703] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:48.473 [2024-09-30 14:09:52.991805] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.473 [2024-09-30 14:09:52.991837] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:48.473 [2024-09-30 14:09:52.991870] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.473 [2024-09-30 14:09:52.993852] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.473 [2024-09-30 14:09:52.993921] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:48.473 BaseBdev3 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 BaseBdev4_malloc 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 true 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.473 [2024-09-30 14:09:53.032192] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:48.473 [2024-09-30 14:09:53.032253] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.473 [2024-09-30 14:09:53.032285] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:48.473 [2024-09-30 14:09:53.032295] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.473 [2024-09-30 14:09:53.034286] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.473 [2024-09-30 14:09:53.034324] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:48.473 BaseBdev4 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:48.473 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.474 [2024-09-30 14:09:53.044229] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:48.474 [2024-09-30 14:09:53.046030] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:48.474 [2024-09-30 14:09:53.046155] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:48.474 [2024-09-30 14:09:53.046213] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:48.474 [2024-09-30 14:09:53.046397] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:10:48.474 [2024-09-30 14:09:53.046412] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:48.474 [2024-09-30 14:09:53.046650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:48.474 [2024-09-30 14:09:53.046792] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:10:48.474 [2024-09-30 14:09:53.046806] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:10:48.474 [2024-09-30 14:09:53.046955] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.474 "name": "raid_bdev1", 00:10:48.474 "uuid": "95a9820d-0e80-4dd8-8c1e-c40f11347192", 00:10:48.474 "strip_size_kb": 0, 00:10:48.474 "state": "online", 00:10:48.474 "raid_level": "raid1", 00:10:48.474 "superblock": true, 00:10:48.474 "num_base_bdevs": 4, 00:10:48.474 "num_base_bdevs_discovered": 4, 00:10:48.474 "num_base_bdevs_operational": 4, 00:10:48.474 "base_bdevs_list": [ 00:10:48.474 { 00:10:48.474 "name": "BaseBdev1", 00:10:48.474 "uuid": "e40938ac-a8bc-58dd-b102-cabf4b3b3842", 00:10:48.474 "is_configured": true, 00:10:48.474 "data_offset": 2048, 00:10:48.474 "data_size": 63488 00:10:48.474 }, 00:10:48.474 { 00:10:48.474 "name": "BaseBdev2", 00:10:48.474 "uuid": "7ee74c4c-f9d8-55f6-ad1b-243ca4822ba2", 00:10:48.474 "is_configured": true, 00:10:48.474 "data_offset": 2048, 00:10:48.474 "data_size": 63488 00:10:48.474 }, 00:10:48.474 { 00:10:48.474 "name": "BaseBdev3", 00:10:48.474 "uuid": "93373e47-7751-59e4-864b-1f2ac9c43efd", 00:10:48.474 "is_configured": true, 00:10:48.474 "data_offset": 2048, 00:10:48.474 "data_size": 63488 00:10:48.474 }, 00:10:48.474 { 00:10:48.474 "name": "BaseBdev4", 00:10:48.474 "uuid": "033b959a-cf4f-5d3c-9229-2dfa1352047c", 00:10:48.474 "is_configured": true, 00:10:48.474 "data_offset": 2048, 00:10:48.474 "data_size": 63488 00:10:48.474 } 00:10:48.474 ] 00:10:48.474 }' 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.474 14:09:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.043 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:49.043 14:09:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:49.043 [2024-09-30 14:09:53.575672] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.983 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.983 "name": "raid_bdev1", 00:10:49.983 "uuid": "95a9820d-0e80-4dd8-8c1e-c40f11347192", 00:10:49.983 "strip_size_kb": 0, 00:10:49.983 "state": "online", 00:10:49.983 "raid_level": "raid1", 00:10:49.983 "superblock": true, 00:10:49.983 "num_base_bdevs": 4, 00:10:49.983 "num_base_bdevs_discovered": 4, 00:10:49.983 "num_base_bdevs_operational": 4, 00:10:49.983 "base_bdevs_list": [ 00:10:49.983 { 00:10:49.983 "name": "BaseBdev1", 00:10:49.983 "uuid": "e40938ac-a8bc-58dd-b102-cabf4b3b3842", 00:10:49.983 "is_configured": true, 00:10:49.983 "data_offset": 2048, 00:10:49.983 "data_size": 63488 00:10:49.983 }, 00:10:49.983 { 00:10:49.983 "name": "BaseBdev2", 00:10:49.983 "uuid": "7ee74c4c-f9d8-55f6-ad1b-243ca4822ba2", 00:10:49.983 "is_configured": true, 00:10:49.983 "data_offset": 2048, 00:10:49.983 "data_size": 63488 00:10:49.983 }, 00:10:49.983 { 00:10:49.983 "name": "BaseBdev3", 00:10:49.983 "uuid": "93373e47-7751-59e4-864b-1f2ac9c43efd", 00:10:49.983 "is_configured": true, 00:10:49.983 "data_offset": 2048, 00:10:49.983 "data_size": 63488 00:10:49.983 }, 00:10:49.983 { 00:10:49.983 "name": "BaseBdev4", 00:10:49.983 "uuid": "033b959a-cf4f-5d3c-9229-2dfa1352047c", 00:10:49.983 "is_configured": true, 00:10:49.983 "data_offset": 2048, 00:10:49.983 "data_size": 63488 00:10:49.983 } 00:10:49.983 ] 00:10:49.984 }' 00:10:49.984 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.984 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.553 [2024-09-30 14:09:54.930078] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.553 [2024-09-30 14:09:54.930207] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:50.553 [2024-09-30 14:09:54.932681] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:50.553 [2024-09-30 14:09:54.932770] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:50.553 [2024-09-30 14:09:54.932896] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:50.553 [2024-09-30 14:09:54.932961] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:10:50.553 { 00:10:50.553 "results": [ 00:10:50.553 { 00:10:50.553 "job": "raid_bdev1", 00:10:50.553 "core_mask": "0x1", 00:10:50.553 "workload": "randrw", 00:10:50.553 "percentage": 50, 00:10:50.553 "status": "finished", 00:10:50.553 "queue_depth": 1, 00:10:50.553 "io_size": 131072, 00:10:50.553 "runtime": 1.355438, 00:10:50.553 "iops": 11942.265157093132, 00:10:50.553 "mibps": 1492.7831446366415, 00:10:50.553 "io_failed": 0, 00:10:50.553 "io_timeout": 0, 00:10:50.553 "avg_latency_us": 81.2928798596534, 00:10:50.553 "min_latency_us": 21.687336244541484, 00:10:50.553 "max_latency_us": 1409.4532751091704 00:10:50.553 } 00:10:50.553 ], 00:10:50.553 "core_count": 1 00:10:50.553 } 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 86777 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 86777 ']' 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 86777 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86777 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:50.553 killing process with pid 86777 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86777' 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 86777 00:10:50.553 [2024-09-30 14:09:54.967117] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:50.553 14:09:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 86777 00:10:50.553 [2024-09-30 14:09:55.001786] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.T1Fb57H9Pk 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:50.813 ************************************ 00:10:50.813 END TEST raid_read_error_test 00:10:50.813 ************************************ 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:50.813 00:10:50.813 real 0m3.321s 00:10:50.813 user 0m4.123s 00:10:50.813 sys 0m0.574s 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:50.813 14:09:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.813 14:09:55 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:10:50.813 14:09:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:50.813 14:09:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:50.813 14:09:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:50.813 ************************************ 00:10:50.813 START TEST raid_write_error_test 00:10:50.813 ************************************ 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 write 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:50.813 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.uOzNPcYSXt 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=86912 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 86912 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 86912 ']' 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:50.814 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:50.814 14:09:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.814 [2024-09-30 14:09:55.439694] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:50.814 [2024-09-30 14:09:55.439931] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86912 ] 00:10:51.074 [2024-09-30 14:09:55.575651] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:51.074 [2024-09-30 14:09:55.605197] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:51.074 [2024-09-30 14:09:55.651071] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:51.074 [2024-09-30 14:09:55.692919] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:51.074 [2024-09-30 14:09:55.692962] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.642 BaseBdev1_malloc 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.642 true 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.642 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 [2024-09-30 14:09:56.299539] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:51.902 [2024-09-30 14:09:56.299602] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.902 [2024-09-30 14:09:56.299628] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:51.902 [2024-09-30 14:09:56.299642] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.902 [2024-09-30 14:09:56.301687] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.902 [2024-09-30 14:09:56.301723] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:51.902 BaseBdev1 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 BaseBdev2_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 true 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 [2024-09-30 14:09:56.360828] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:51.902 [2024-09-30 14:09:56.360902] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.902 [2024-09-30 14:09:56.360926] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:51.902 [2024-09-30 14:09:56.360942] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.902 [2024-09-30 14:09:56.363858] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.902 [2024-09-30 14:09:56.363916] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:51.902 BaseBdev2 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 BaseBdev3_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 true 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.902 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.902 [2024-09-30 14:09:56.401863] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:51.902 [2024-09-30 14:09:56.401919] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.902 [2024-09-30 14:09:56.401937] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:51.903 [2024-09-30 14:09:56.401947] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.903 [2024-09-30 14:09:56.404069] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.903 [2024-09-30 14:09:56.404108] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:51.903 BaseBdev3 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.903 BaseBdev4_malloc 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.903 true 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.903 [2024-09-30 14:09:56.442143] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:51.903 [2024-09-30 14:09:56.442197] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.903 [2024-09-30 14:09:56.442215] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:51.903 [2024-09-30 14:09:56.442226] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.903 [2024-09-30 14:09:56.444216] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.903 [2024-09-30 14:09:56.444336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:51.903 BaseBdev4 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.903 [2024-09-30 14:09:56.454178] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:51.903 [2024-09-30 14:09:56.455993] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:51.903 [2024-09-30 14:09:56.456064] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:51.903 [2024-09-30 14:09:56.456124] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:51.903 [2024-09-30 14:09:56.456299] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:10:51.903 [2024-09-30 14:09:56.456317] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:51.903 [2024-09-30 14:09:56.456548] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:51.903 [2024-09-30 14:09:56.456681] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:10:51.903 [2024-09-30 14:09:56.456692] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:10:51.903 [2024-09-30 14:09:56.456814] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.903 "name": "raid_bdev1", 00:10:51.903 "uuid": "b0d88a26-1ddb-4364-b77f-89b576308195", 00:10:51.903 "strip_size_kb": 0, 00:10:51.903 "state": "online", 00:10:51.903 "raid_level": "raid1", 00:10:51.903 "superblock": true, 00:10:51.903 "num_base_bdevs": 4, 00:10:51.903 "num_base_bdevs_discovered": 4, 00:10:51.903 "num_base_bdevs_operational": 4, 00:10:51.903 "base_bdevs_list": [ 00:10:51.903 { 00:10:51.903 "name": "BaseBdev1", 00:10:51.903 "uuid": "cdcdae1d-4ff4-5874-98bd-ec31f67412d6", 00:10:51.903 "is_configured": true, 00:10:51.903 "data_offset": 2048, 00:10:51.903 "data_size": 63488 00:10:51.903 }, 00:10:51.903 { 00:10:51.903 "name": "BaseBdev2", 00:10:51.903 "uuid": "94cb8dab-b196-591b-902a-ee4d70f6c9a0", 00:10:51.903 "is_configured": true, 00:10:51.903 "data_offset": 2048, 00:10:51.903 "data_size": 63488 00:10:51.903 }, 00:10:51.903 { 00:10:51.903 "name": "BaseBdev3", 00:10:51.903 "uuid": "5d805fd7-0454-5a88-94b6-f3075fb85360", 00:10:51.903 "is_configured": true, 00:10:51.903 "data_offset": 2048, 00:10:51.903 "data_size": 63488 00:10:51.903 }, 00:10:51.903 { 00:10:51.903 "name": "BaseBdev4", 00:10:51.903 "uuid": "2a7977a3-ad83-54d3-b579-54206e2c3398", 00:10:51.903 "is_configured": true, 00:10:51.903 "data_offset": 2048, 00:10:51.903 "data_size": 63488 00:10:51.903 } 00:10:51.903 ] 00:10:51.903 }' 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.903 14:09:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.480 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:52.480 14:09:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:52.480 [2024-09-30 14:09:56.965723] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.422 [2024-09-30 14:09:57.889385] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:53.422 [2024-09-30 14:09:57.889465] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:53.422 [2024-09-30 14:09:57.889702] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006150 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.422 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.423 "name": "raid_bdev1", 00:10:53.423 "uuid": "b0d88a26-1ddb-4364-b77f-89b576308195", 00:10:53.423 "strip_size_kb": 0, 00:10:53.423 "state": "online", 00:10:53.423 "raid_level": "raid1", 00:10:53.423 "superblock": true, 00:10:53.423 "num_base_bdevs": 4, 00:10:53.423 "num_base_bdevs_discovered": 3, 00:10:53.423 "num_base_bdevs_operational": 3, 00:10:53.423 "base_bdevs_list": [ 00:10:53.423 { 00:10:53.423 "name": null, 00:10:53.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.423 "is_configured": false, 00:10:53.423 "data_offset": 0, 00:10:53.423 "data_size": 63488 00:10:53.423 }, 00:10:53.423 { 00:10:53.423 "name": "BaseBdev2", 00:10:53.423 "uuid": "94cb8dab-b196-591b-902a-ee4d70f6c9a0", 00:10:53.423 "is_configured": true, 00:10:53.423 "data_offset": 2048, 00:10:53.423 "data_size": 63488 00:10:53.423 }, 00:10:53.423 { 00:10:53.423 "name": "BaseBdev3", 00:10:53.423 "uuid": "5d805fd7-0454-5a88-94b6-f3075fb85360", 00:10:53.423 "is_configured": true, 00:10:53.423 "data_offset": 2048, 00:10:53.423 "data_size": 63488 00:10:53.423 }, 00:10:53.423 { 00:10:53.423 "name": "BaseBdev4", 00:10:53.423 "uuid": "2a7977a3-ad83-54d3-b579-54206e2c3398", 00:10:53.423 "is_configured": true, 00:10:53.423 "data_offset": 2048, 00:10:53.423 "data_size": 63488 00:10:53.423 } 00:10:53.423 ] 00:10:53.423 }' 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.423 14:09:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.682 [2024-09-30 14:09:58.296025] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:53.682 [2024-09-30 14:09:58.296155] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:53.682 [2024-09-30 14:09:58.298634] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:53.682 [2024-09-30 14:09:58.298728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.682 [2024-09-30 14:09:58.298845] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:53.682 [2024-09-30 14:09:58.298980] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:10:53.682 { 00:10:53.682 "results": [ 00:10:53.682 { 00:10:53.682 "job": "raid_bdev1", 00:10:53.682 "core_mask": "0x1", 00:10:53.682 "workload": "randrw", 00:10:53.682 "percentage": 50, 00:10:53.682 "status": "finished", 00:10:53.682 "queue_depth": 1, 00:10:53.682 "io_size": 131072, 00:10:53.682 "runtime": 1.330898, 00:10:53.682 "iops": 12403.655276362275, 00:10:53.682 "mibps": 1550.4569095452844, 00:10:53.682 "io_failed": 0, 00:10:53.682 "io_timeout": 0, 00:10:53.682 "avg_latency_us": 78.04690170069719, 00:10:53.682 "min_latency_us": 22.358078602620086, 00:10:53.682 "max_latency_us": 1423.7624454148472 00:10:53.682 } 00:10:53.682 ], 00:10:53.682 "core_count": 1 00:10:53.682 } 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 86912 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 86912 ']' 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 86912 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86912 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:53.682 killing process with pid 86912 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86912' 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 86912 00:10:53.682 [2024-09-30 14:09:58.337410] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:53.682 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 86912 00:10:53.942 [2024-09-30 14:09:58.372249] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.uOzNPcYSXt 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:54.201 ************************************ 00:10:54.201 END TEST raid_write_error_test 00:10:54.201 ************************************ 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:54.201 00:10:54.201 real 0m3.300s 00:10:54.201 user 0m4.047s 00:10:54.201 sys 0m0.605s 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:54.201 14:09:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.201 14:09:58 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:10:54.201 14:09:58 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:10:54.201 14:09:58 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:10:54.201 14:09:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:10:54.201 14:09:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:54.201 14:09:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:54.201 ************************************ 00:10:54.201 START TEST raid_rebuild_test 00:10:54.201 ************************************ 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false false true 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=87039 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 87039 00:10:54.201 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 87039 ']' 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:54.201 14:09:58 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.201 I/O size of 3145728 is greater than zero copy threshold (65536). 00:10:54.201 Zero copy mechanism will not be used. 00:10:54.201 [2024-09-30 14:09:58.794677] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:10:54.201 [2024-09-30 14:09:58.794793] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87039 ] 00:10:54.461 [2024-09-30 14:09:58.925337] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:10:54.461 [2024-09-30 14:09:58.943824] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:54.461 [2024-09-30 14:09:58.988675] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:54.461 [2024-09-30 14:09:59.029606] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:54.461 [2024-09-30 14:09:59.029642] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.033 BaseBdev1_malloc 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.033 [2024-09-30 14:09:59.647134] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:55.033 [2024-09-30 14:09:59.647276] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.033 [2024-09-30 14:09:59.647313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:55.033 [2024-09-30 14:09:59.647337] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.033 [2024-09-30 14:09:59.649390] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.033 [2024-09-30 14:09:59.649427] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:55.033 BaseBdev1 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.033 BaseBdev2_malloc 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.033 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.033 [2024-09-30 14:09:59.684192] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:55.033 [2024-09-30 14:09:59.684319] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.033 [2024-09-30 14:09:59.684348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:55.033 [2024-09-30 14:09:59.684362] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.033 [2024-09-30 14:09:59.686476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.033 [2024-09-30 14:09:59.686518] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:55.293 BaseBdev2 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.293 spare_malloc 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.293 spare_delay 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.293 [2024-09-30 14:09:59.724491] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:55.293 [2024-09-30 14:09:59.724557] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.293 [2024-09-30 14:09:59.724590] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:10:55.293 [2024-09-30 14:09:59.724600] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.293 [2024-09-30 14:09:59.726605] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.293 [2024-09-30 14:09:59.726693] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:55.293 spare 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.293 [2024-09-30 14:09:59.736532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:55.293 [2024-09-30 14:09:59.738333] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:55.293 [2024-09-30 14:09:59.738418] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:10:55.293 [2024-09-30 14:09:59.738430] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:55.293 [2024-09-30 14:09:59.738686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:55.293 [2024-09-30 14:09:59.738801] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:10:55.293 [2024-09-30 14:09:59.738811] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:10:55.293 [2024-09-30 14:09:59.738937] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:55.293 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.294 "name": "raid_bdev1", 00:10:55.294 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:10:55.294 "strip_size_kb": 0, 00:10:55.294 "state": "online", 00:10:55.294 "raid_level": "raid1", 00:10:55.294 "superblock": false, 00:10:55.294 "num_base_bdevs": 2, 00:10:55.294 "num_base_bdevs_discovered": 2, 00:10:55.294 "num_base_bdevs_operational": 2, 00:10:55.294 "base_bdevs_list": [ 00:10:55.294 { 00:10:55.294 "name": "BaseBdev1", 00:10:55.294 "uuid": "2b6aa46f-208f-5706-b9ff-fe59c192d11e", 00:10:55.294 "is_configured": true, 00:10:55.294 "data_offset": 0, 00:10:55.294 "data_size": 65536 00:10:55.294 }, 00:10:55.294 { 00:10:55.294 "name": "BaseBdev2", 00:10:55.294 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:10:55.294 "is_configured": true, 00:10:55.294 "data_offset": 0, 00:10:55.294 "data_size": 65536 00:10:55.294 } 00:10:55.294 ] 00:10:55.294 }' 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.294 14:09:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.554 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:55.554 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:55.554 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.554 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.554 [2024-09-30 14:10:00.180024] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:55.554 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:55.814 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:10:55.814 [2024-09-30 14:10:00.459396] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:10:56.074 /dev/nbd0 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:56.074 1+0 records in 00:10:56.074 1+0 records out 00:10:56.074 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00040437 s, 10.1 MB/s 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:10:56.074 14:10:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:00.268 65536+0 records in 00:11:00.268 65536+0 records out 00:11:00.268 33554432 bytes (34 MB, 32 MiB) copied, 4.15573 s, 8.1 MB/s 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:00.268 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:00.268 [2024-09-30 14:10:04.906472] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.528 [2024-09-30 14:10:04.938379] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:00.528 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.529 "name": "raid_bdev1", 00:11:00.529 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:00.529 "strip_size_kb": 0, 00:11:00.529 "state": "online", 00:11:00.529 "raid_level": "raid1", 00:11:00.529 "superblock": false, 00:11:00.529 "num_base_bdevs": 2, 00:11:00.529 "num_base_bdevs_discovered": 1, 00:11:00.529 "num_base_bdevs_operational": 1, 00:11:00.529 "base_bdevs_list": [ 00:11:00.529 { 00:11:00.529 "name": null, 00:11:00.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.529 "is_configured": false, 00:11:00.529 "data_offset": 0, 00:11:00.529 "data_size": 65536 00:11:00.529 }, 00:11:00.529 { 00:11:00.529 "name": "BaseBdev2", 00:11:00.529 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:00.529 "is_configured": true, 00:11:00.529 "data_offset": 0, 00:11:00.529 "data_size": 65536 00:11:00.529 } 00:11:00.529 ] 00:11:00.529 }' 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.529 14:10:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.789 14:10:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:00.789 14:10:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.789 14:10:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.789 [2024-09-30 14:10:05.357648] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:00.789 [2024-09-30 14:10:05.361915] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:11:00.789 14:10:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.789 14:10:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:00.789 [2024-09-30 14:10:05.363947] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.728 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:01.990 "name": "raid_bdev1", 00:11:01.990 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:01.990 "strip_size_kb": 0, 00:11:01.990 "state": "online", 00:11:01.990 "raid_level": "raid1", 00:11:01.990 "superblock": false, 00:11:01.990 "num_base_bdevs": 2, 00:11:01.990 "num_base_bdevs_discovered": 2, 00:11:01.990 "num_base_bdevs_operational": 2, 00:11:01.990 "process": { 00:11:01.990 "type": "rebuild", 00:11:01.990 "target": "spare", 00:11:01.990 "progress": { 00:11:01.990 "blocks": 20480, 00:11:01.990 "percent": 31 00:11:01.990 } 00:11:01.990 }, 00:11:01.990 "base_bdevs_list": [ 00:11:01.990 { 00:11:01.990 "name": "spare", 00:11:01.990 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:01.990 "is_configured": true, 00:11:01.990 "data_offset": 0, 00:11:01.990 "data_size": 65536 00:11:01.990 }, 00:11:01.990 { 00:11:01.990 "name": "BaseBdev2", 00:11:01.990 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:01.990 "is_configured": true, 00:11:01.990 "data_offset": 0, 00:11:01.990 "data_size": 65536 00:11:01.990 } 00:11:01.990 ] 00:11:01.990 }' 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.990 [2024-09-30 14:10:06.508992] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:01.990 [2024-09-30 14:10:06.568895] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:01.990 [2024-09-30 14:10:06.569043] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:01.990 [2024-09-30 14:10:06.569061] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:01.990 [2024-09-30 14:10:06.569071] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.990 "name": "raid_bdev1", 00:11:01.990 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:01.990 "strip_size_kb": 0, 00:11:01.990 "state": "online", 00:11:01.990 "raid_level": "raid1", 00:11:01.990 "superblock": false, 00:11:01.990 "num_base_bdevs": 2, 00:11:01.990 "num_base_bdevs_discovered": 1, 00:11:01.990 "num_base_bdevs_operational": 1, 00:11:01.990 "base_bdevs_list": [ 00:11:01.990 { 00:11:01.990 "name": null, 00:11:01.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.990 "is_configured": false, 00:11:01.990 "data_offset": 0, 00:11:01.990 "data_size": 65536 00:11:01.990 }, 00:11:01.990 { 00:11:01.990 "name": "BaseBdev2", 00:11:01.990 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:01.990 "is_configured": true, 00:11:01.990 "data_offset": 0, 00:11:01.990 "data_size": 65536 00:11:01.990 } 00:11:01.990 ] 00:11:01.990 }' 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.990 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.560 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:02.560 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:02.560 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:02.560 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:02.561 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:02.561 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.561 14:10:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.561 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.561 14:10:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:02.561 "name": "raid_bdev1", 00:11:02.561 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:02.561 "strip_size_kb": 0, 00:11:02.561 "state": "online", 00:11:02.561 "raid_level": "raid1", 00:11:02.561 "superblock": false, 00:11:02.561 "num_base_bdevs": 2, 00:11:02.561 "num_base_bdevs_discovered": 1, 00:11:02.561 "num_base_bdevs_operational": 1, 00:11:02.561 "base_bdevs_list": [ 00:11:02.561 { 00:11:02.561 "name": null, 00:11:02.561 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.561 "is_configured": false, 00:11:02.561 "data_offset": 0, 00:11:02.561 "data_size": 65536 00:11:02.561 }, 00:11:02.561 { 00:11:02.561 "name": "BaseBdev2", 00:11:02.561 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:02.561 "is_configured": true, 00:11:02.561 "data_offset": 0, 00:11:02.561 "data_size": 65536 00:11:02.561 } 00:11:02.561 ] 00:11:02.561 }' 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.561 [2024-09-30 14:10:07.116635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:02.561 [2024-09-30 14:10:07.120749] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09d70 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.561 14:10:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:02.561 [2024-09-30 14:10:07.122503] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.499 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.759 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:03.760 "name": "raid_bdev1", 00:11:03.760 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:03.760 "strip_size_kb": 0, 00:11:03.760 "state": "online", 00:11:03.760 "raid_level": "raid1", 00:11:03.760 "superblock": false, 00:11:03.760 "num_base_bdevs": 2, 00:11:03.760 "num_base_bdevs_discovered": 2, 00:11:03.760 "num_base_bdevs_operational": 2, 00:11:03.760 "process": { 00:11:03.760 "type": "rebuild", 00:11:03.760 "target": "spare", 00:11:03.760 "progress": { 00:11:03.760 "blocks": 20480, 00:11:03.760 "percent": 31 00:11:03.760 } 00:11:03.760 }, 00:11:03.760 "base_bdevs_list": [ 00:11:03.760 { 00:11:03.760 "name": "spare", 00:11:03.760 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:03.760 "is_configured": true, 00:11:03.760 "data_offset": 0, 00:11:03.760 "data_size": 65536 00:11:03.760 }, 00:11:03.760 { 00:11:03.760 "name": "BaseBdev2", 00:11:03.760 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:03.760 "is_configured": true, 00:11:03.760 "data_offset": 0, 00:11:03.760 "data_size": 65536 00:11:03.760 } 00:11:03.760 ] 00:11:03.760 }' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=289 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:03.760 "name": "raid_bdev1", 00:11:03.760 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:03.760 "strip_size_kb": 0, 00:11:03.760 "state": "online", 00:11:03.760 "raid_level": "raid1", 00:11:03.760 "superblock": false, 00:11:03.760 "num_base_bdevs": 2, 00:11:03.760 "num_base_bdevs_discovered": 2, 00:11:03.760 "num_base_bdevs_operational": 2, 00:11:03.760 "process": { 00:11:03.760 "type": "rebuild", 00:11:03.760 "target": "spare", 00:11:03.760 "progress": { 00:11:03.760 "blocks": 22528, 00:11:03.760 "percent": 34 00:11:03.760 } 00:11:03.760 }, 00:11:03.760 "base_bdevs_list": [ 00:11:03.760 { 00:11:03.760 "name": "spare", 00:11:03.760 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:03.760 "is_configured": true, 00:11:03.760 "data_offset": 0, 00:11:03.760 "data_size": 65536 00:11:03.760 }, 00:11:03.760 { 00:11:03.760 "name": "BaseBdev2", 00:11:03.760 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:03.760 "is_configured": true, 00:11:03.760 "data_offset": 0, 00:11:03.760 "data_size": 65536 00:11:03.760 } 00:11:03.760 ] 00:11:03.760 }' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:03.760 14:10:08 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:05.141 "name": "raid_bdev1", 00:11:05.141 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:05.141 "strip_size_kb": 0, 00:11:05.141 "state": "online", 00:11:05.141 "raid_level": "raid1", 00:11:05.141 "superblock": false, 00:11:05.141 "num_base_bdevs": 2, 00:11:05.141 "num_base_bdevs_discovered": 2, 00:11:05.141 "num_base_bdevs_operational": 2, 00:11:05.141 "process": { 00:11:05.141 "type": "rebuild", 00:11:05.141 "target": "spare", 00:11:05.141 "progress": { 00:11:05.141 "blocks": 45056, 00:11:05.141 "percent": 68 00:11:05.141 } 00:11:05.141 }, 00:11:05.141 "base_bdevs_list": [ 00:11:05.141 { 00:11:05.141 "name": "spare", 00:11:05.141 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:05.141 "is_configured": true, 00:11:05.141 "data_offset": 0, 00:11:05.141 "data_size": 65536 00:11:05.141 }, 00:11:05.141 { 00:11:05.141 "name": "BaseBdev2", 00:11:05.141 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:05.141 "is_configured": true, 00:11:05.141 "data_offset": 0, 00:11:05.141 "data_size": 65536 00:11:05.141 } 00:11:05.141 ] 00:11:05.141 }' 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:05.141 14:10:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:05.710 [2024-09-30 14:10:10.333783] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:05.710 [2024-09-30 14:10:10.333853] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:05.710 [2024-09-30 14:10:10.333910] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:05.970 "name": "raid_bdev1", 00:11:05.970 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:05.970 "strip_size_kb": 0, 00:11:05.970 "state": "online", 00:11:05.970 "raid_level": "raid1", 00:11:05.970 "superblock": false, 00:11:05.970 "num_base_bdevs": 2, 00:11:05.970 "num_base_bdevs_discovered": 2, 00:11:05.970 "num_base_bdevs_operational": 2, 00:11:05.970 "base_bdevs_list": [ 00:11:05.970 { 00:11:05.970 "name": "spare", 00:11:05.970 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:05.970 "is_configured": true, 00:11:05.970 "data_offset": 0, 00:11:05.970 "data_size": 65536 00:11:05.970 }, 00:11:05.970 { 00:11:05.970 "name": "BaseBdev2", 00:11:05.970 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:05.970 "is_configured": true, 00:11:05.970 "data_offset": 0, 00:11:05.970 "data_size": 65536 00:11:05.970 } 00:11:05.970 ] 00:11:05.970 }' 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.970 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:06.243 "name": "raid_bdev1", 00:11:06.243 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:06.243 "strip_size_kb": 0, 00:11:06.243 "state": "online", 00:11:06.243 "raid_level": "raid1", 00:11:06.243 "superblock": false, 00:11:06.243 "num_base_bdevs": 2, 00:11:06.243 "num_base_bdevs_discovered": 2, 00:11:06.243 "num_base_bdevs_operational": 2, 00:11:06.243 "base_bdevs_list": [ 00:11:06.243 { 00:11:06.243 "name": "spare", 00:11:06.243 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:06.243 "is_configured": true, 00:11:06.243 "data_offset": 0, 00:11:06.243 "data_size": 65536 00:11:06.243 }, 00:11:06.243 { 00:11:06.243 "name": "BaseBdev2", 00:11:06.243 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:06.243 "is_configured": true, 00:11:06.243 "data_offset": 0, 00:11:06.243 "data_size": 65536 00:11:06.243 } 00:11:06.243 ] 00:11:06.243 }' 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.243 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.243 "name": "raid_bdev1", 00:11:06.243 "uuid": "67e65d1f-e852-4d0c-9b73-b6591646fca2", 00:11:06.243 "strip_size_kb": 0, 00:11:06.244 "state": "online", 00:11:06.244 "raid_level": "raid1", 00:11:06.244 "superblock": false, 00:11:06.244 "num_base_bdevs": 2, 00:11:06.244 "num_base_bdevs_discovered": 2, 00:11:06.244 "num_base_bdevs_operational": 2, 00:11:06.244 "base_bdevs_list": [ 00:11:06.244 { 00:11:06.244 "name": "spare", 00:11:06.244 "uuid": "4c1b7b32-d599-5d1e-9a8f-1786e76310f9", 00:11:06.244 "is_configured": true, 00:11:06.244 "data_offset": 0, 00:11:06.244 "data_size": 65536 00:11:06.244 }, 00:11:06.244 { 00:11:06.244 "name": "BaseBdev2", 00:11:06.244 "uuid": "3397b088-ecb2-5709-9cba-1028ebb6264c", 00:11:06.244 "is_configured": true, 00:11:06.244 "data_offset": 0, 00:11:06.244 "data_size": 65536 00:11:06.244 } 00:11:06.244 ] 00:11:06.244 }' 00:11:06.244 14:10:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.244 14:10:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.521 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.522 [2024-09-30 14:10:11.156765] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:06.522 [2024-09-30 14:10:11.156880] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:06.522 [2024-09-30 14:10:11.156989] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:06.522 [2024-09-30 14:10:11.157084] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:06.522 [2024-09-30 14:10:11.157168] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.522 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:06.780 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:06.780 /dev/nbd0 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:07.038 1+0 records in 00:11:07.038 1+0 records out 00:11:07.038 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000256004 s, 16.0 MB/s 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:07.038 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:07.038 /dev/nbd1 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:07.039 1+0 records in 00:11:07.039 1+0 records out 00:11:07.039 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000259459 s, 15.8 MB/s 00:11:07.039 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:07.298 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:07.557 14:10:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 87039 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 87039 ']' 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 87039 00:11:07.816 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87039 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:07.817 killing process with pid 87039 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87039' 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 87039 00:11:07.817 Received shutdown signal, test time was about 60.000000 seconds 00:11:07.817 00:11:07.817 Latency(us) 00:11:07.817 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:07.817 =================================================================================================================== 00:11:07.817 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:07.817 [2024-09-30 14:10:12.300842] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:07.817 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 87039 00:11:07.817 [2024-09-30 14:10:12.330927] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:08.077 00:11:08.077 real 0m13.870s 00:11:08.077 user 0m15.638s 00:11:08.077 sys 0m3.047s 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.077 ************************************ 00:11:08.077 END TEST raid_rebuild_test 00:11:08.077 ************************************ 00:11:08.077 14:10:12 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:08.077 14:10:12 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:08.077 14:10:12 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:08.077 14:10:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:08.077 ************************************ 00:11:08.077 START TEST raid_rebuild_test_sb 00:11:08.077 ************************************ 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=87445 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 87445 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 87445 ']' 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:08.077 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:08.077 14:10:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.337 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:08.337 Zero copy mechanism will not be used. 00:11:08.337 [2024-09-30 14:10:12.741207] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:11:08.337 [2024-09-30 14:10:12.741323] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87445 ] 00:11:08.337 [2024-09-30 14:10:12.871589] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:11:08.337 [2024-09-30 14:10:12.900563] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:08.337 [2024-09-30 14:10:12.945116] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:08.337 [2024-09-30 14:10:12.986182] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:08.337 [2024-09-30 14:10:12.986220] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 BaseBdev1_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 [2024-09-30 14:10:13.667859] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:09.274 [2024-09-30 14:10:13.667929] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.274 [2024-09-30 14:10:13.667955] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:09.274 [2024-09-30 14:10:13.667972] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.274 [2024-09-30 14:10:13.670009] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.274 [2024-09-30 14:10:13.670043] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:09.274 BaseBdev1 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 BaseBdev2_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 [2024-09-30 14:10:13.704018] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:09.274 [2024-09-30 14:10:13.704075] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.274 [2024-09-30 14:10:13.704095] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:09.274 [2024-09-30 14:10:13.704107] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.274 [2024-09-30 14:10:13.706433] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.274 [2024-09-30 14:10:13.706470] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:09.274 BaseBdev2 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 spare_malloc 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 spare_delay 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 [2024-09-30 14:10:13.744307] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:09.274 [2024-09-30 14:10:13.744358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.274 [2024-09-30 14:10:13.744390] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:09.274 [2024-09-30 14:10:13.744400] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.274 [2024-09-30 14:10:13.746418] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.274 [2024-09-30 14:10:13.746456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:09.274 spare 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 [2024-09-30 14:10:13.756341] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:09.274 [2024-09-30 14:10:13.758075] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:09.274 [2024-09-30 14:10:13.758215] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:11:09.274 [2024-09-30 14:10:13.758236] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:09.274 [2024-09-30 14:10:13.758468] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:09.274 [2024-09-30 14:10:13.758615] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:11:09.274 [2024-09-30 14:10:13.758632] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:11:09.274 [2024-09-30 14:10:13.758749] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.274 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:09.274 "name": "raid_bdev1", 00:11:09.275 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:09.275 "strip_size_kb": 0, 00:11:09.275 "state": "online", 00:11:09.275 "raid_level": "raid1", 00:11:09.275 "superblock": true, 00:11:09.275 "num_base_bdevs": 2, 00:11:09.275 "num_base_bdevs_discovered": 2, 00:11:09.275 "num_base_bdevs_operational": 2, 00:11:09.275 "base_bdevs_list": [ 00:11:09.275 { 00:11:09.275 "name": "BaseBdev1", 00:11:09.275 "uuid": "93ea7a24-3722-5db8-b85e-ee118e7c3859", 00:11:09.275 "is_configured": true, 00:11:09.275 "data_offset": 2048, 00:11:09.275 "data_size": 63488 00:11:09.275 }, 00:11:09.275 { 00:11:09.275 "name": "BaseBdev2", 00:11:09.275 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:09.275 "is_configured": true, 00:11:09.275 "data_offset": 2048, 00:11:09.275 "data_size": 63488 00:11:09.275 } 00:11:09.275 ] 00:11:09.275 }' 00:11:09.275 14:10:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:09.275 14:10:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.844 [2024-09-30 14:10:14.211860] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:09.844 [2024-09-30 14:10:14.463199] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:11:09.844 /dev/nbd0 00:11:09.844 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:10.103 1+0 records in 00:11:10.103 1+0 records out 00:11:10.103 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000302205 s, 13.6 MB/s 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:10.103 14:10:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:14.295 63488+0 records in 00:11:14.295 63488+0 records out 00:11:14.295 32505856 bytes (33 MB, 31 MiB) copied, 4.1068 s, 7.9 MB/s 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:14.295 [2024-09-30 14:10:18.847038] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.295 [2024-09-30 14:10:18.864403] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.295 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.295 "name": "raid_bdev1", 00:11:14.296 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:14.296 "strip_size_kb": 0, 00:11:14.296 "state": "online", 00:11:14.296 "raid_level": "raid1", 00:11:14.296 "superblock": true, 00:11:14.296 "num_base_bdevs": 2, 00:11:14.296 "num_base_bdevs_discovered": 1, 00:11:14.296 "num_base_bdevs_operational": 1, 00:11:14.296 "base_bdevs_list": [ 00:11:14.296 { 00:11:14.296 "name": null, 00:11:14.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.296 "is_configured": false, 00:11:14.296 "data_offset": 0, 00:11:14.296 "data_size": 63488 00:11:14.296 }, 00:11:14.296 { 00:11:14.296 "name": "BaseBdev2", 00:11:14.296 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:14.296 "is_configured": true, 00:11:14.296 "data_offset": 2048, 00:11:14.296 "data_size": 63488 00:11:14.296 } 00:11:14.296 ] 00:11:14.296 }' 00:11:14.296 14:10:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.296 14:10:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.862 14:10:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:14.862 14:10:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.862 14:10:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.862 [2024-09-30 14:10:19.315635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:14.862 [2024-09-30 14:10:19.319688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:11:14.862 14:10:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.862 14:10:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:14.862 [2024-09-30 14:10:19.321511] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:15.800 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:15.801 "name": "raid_bdev1", 00:11:15.801 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:15.801 "strip_size_kb": 0, 00:11:15.801 "state": "online", 00:11:15.801 "raid_level": "raid1", 00:11:15.801 "superblock": true, 00:11:15.801 "num_base_bdevs": 2, 00:11:15.801 "num_base_bdevs_discovered": 2, 00:11:15.801 "num_base_bdevs_operational": 2, 00:11:15.801 "process": { 00:11:15.801 "type": "rebuild", 00:11:15.801 "target": "spare", 00:11:15.801 "progress": { 00:11:15.801 "blocks": 20480, 00:11:15.801 "percent": 32 00:11:15.801 } 00:11:15.801 }, 00:11:15.801 "base_bdevs_list": [ 00:11:15.801 { 00:11:15.801 "name": "spare", 00:11:15.801 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:15.801 "is_configured": true, 00:11:15.801 "data_offset": 2048, 00:11:15.801 "data_size": 63488 00:11:15.801 }, 00:11:15.801 { 00:11:15.801 "name": "BaseBdev2", 00:11:15.801 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:15.801 "is_configured": true, 00:11:15.801 "data_offset": 2048, 00:11:15.801 "data_size": 63488 00:11:15.801 } 00:11:15.801 ] 00:11:15.801 }' 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:15.801 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.061 [2024-09-30 14:10:20.486658] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:16.061 [2024-09-30 14:10:20.525924] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:16.061 [2024-09-30 14:10:20.525983] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:16.061 [2024-09-30 14:10:20.526013] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:16.061 [2024-09-30 14:10:20.526033] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.061 "name": "raid_bdev1", 00:11:16.061 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:16.061 "strip_size_kb": 0, 00:11:16.061 "state": "online", 00:11:16.061 "raid_level": "raid1", 00:11:16.061 "superblock": true, 00:11:16.061 "num_base_bdevs": 2, 00:11:16.061 "num_base_bdevs_discovered": 1, 00:11:16.061 "num_base_bdevs_operational": 1, 00:11:16.061 "base_bdevs_list": [ 00:11:16.061 { 00:11:16.061 "name": null, 00:11:16.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:16.061 "is_configured": false, 00:11:16.061 "data_offset": 0, 00:11:16.061 "data_size": 63488 00:11:16.061 }, 00:11:16.061 { 00:11:16.061 "name": "BaseBdev2", 00:11:16.061 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:16.061 "is_configured": true, 00:11:16.061 "data_offset": 2048, 00:11:16.061 "data_size": 63488 00:11:16.061 } 00:11:16.061 ] 00:11:16.061 }' 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.061 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:16.321 "name": "raid_bdev1", 00:11:16.321 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:16.321 "strip_size_kb": 0, 00:11:16.321 "state": "online", 00:11:16.321 "raid_level": "raid1", 00:11:16.321 "superblock": true, 00:11:16.321 "num_base_bdevs": 2, 00:11:16.321 "num_base_bdevs_discovered": 1, 00:11:16.321 "num_base_bdevs_operational": 1, 00:11:16.321 "base_bdevs_list": [ 00:11:16.321 { 00:11:16.321 "name": null, 00:11:16.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:16.321 "is_configured": false, 00:11:16.321 "data_offset": 0, 00:11:16.321 "data_size": 63488 00:11:16.321 }, 00:11:16.321 { 00:11:16.321 "name": "BaseBdev2", 00:11:16.321 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:16.321 "is_configured": true, 00:11:16.321 "data_offset": 2048, 00:11:16.321 "data_size": 63488 00:11:16.321 } 00:11:16.321 ] 00:11:16.321 }' 00:11:16.321 14:10:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.580 [2024-09-30 14:10:21.065536] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:16.580 [2024-09-30 14:10:21.069584] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3500 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.580 14:10:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:16.580 [2024-09-30 14:10:21.071336] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.519 "name": "raid_bdev1", 00:11:17.519 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:17.519 "strip_size_kb": 0, 00:11:17.519 "state": "online", 00:11:17.519 "raid_level": "raid1", 00:11:17.519 "superblock": true, 00:11:17.519 "num_base_bdevs": 2, 00:11:17.519 "num_base_bdevs_discovered": 2, 00:11:17.519 "num_base_bdevs_operational": 2, 00:11:17.519 "process": { 00:11:17.519 "type": "rebuild", 00:11:17.519 "target": "spare", 00:11:17.519 "progress": { 00:11:17.519 "blocks": 20480, 00:11:17.519 "percent": 32 00:11:17.519 } 00:11:17.519 }, 00:11:17.519 "base_bdevs_list": [ 00:11:17.519 { 00:11:17.519 "name": "spare", 00:11:17.519 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:17.519 "is_configured": true, 00:11:17.519 "data_offset": 2048, 00:11:17.519 "data_size": 63488 00:11:17.519 }, 00:11:17.519 { 00:11:17.519 "name": "BaseBdev2", 00:11:17.519 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:17.519 "is_configured": true, 00:11:17.519 "data_offset": 2048, 00:11:17.519 "data_size": 63488 00:11:17.519 } 00:11:17.519 ] 00:11:17.519 }' 00:11:17.519 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:17.779 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=303 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.779 "name": "raid_bdev1", 00:11:17.779 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:17.779 "strip_size_kb": 0, 00:11:17.779 "state": "online", 00:11:17.779 "raid_level": "raid1", 00:11:17.779 "superblock": true, 00:11:17.779 "num_base_bdevs": 2, 00:11:17.779 "num_base_bdevs_discovered": 2, 00:11:17.779 "num_base_bdevs_operational": 2, 00:11:17.779 "process": { 00:11:17.779 "type": "rebuild", 00:11:17.779 "target": "spare", 00:11:17.779 "progress": { 00:11:17.779 "blocks": 22528, 00:11:17.779 "percent": 35 00:11:17.779 } 00:11:17.779 }, 00:11:17.779 "base_bdevs_list": [ 00:11:17.779 { 00:11:17.779 "name": "spare", 00:11:17.779 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:17.779 "is_configured": true, 00:11:17.779 "data_offset": 2048, 00:11:17.779 "data_size": 63488 00:11:17.779 }, 00:11:17.779 { 00:11:17.779 "name": "BaseBdev2", 00:11:17.779 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:17.779 "is_configured": true, 00:11:17.779 "data_offset": 2048, 00:11:17.779 "data_size": 63488 00:11:17.779 } 00:11:17.779 ] 00:11:17.779 }' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:17.779 14:10:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.156 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.156 "name": "raid_bdev1", 00:11:19.156 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:19.156 "strip_size_kb": 0, 00:11:19.156 "state": "online", 00:11:19.156 "raid_level": "raid1", 00:11:19.156 "superblock": true, 00:11:19.156 "num_base_bdevs": 2, 00:11:19.156 "num_base_bdevs_discovered": 2, 00:11:19.156 "num_base_bdevs_operational": 2, 00:11:19.156 "process": { 00:11:19.156 "type": "rebuild", 00:11:19.156 "target": "spare", 00:11:19.157 "progress": { 00:11:19.157 "blocks": 47104, 00:11:19.157 "percent": 74 00:11:19.157 } 00:11:19.157 }, 00:11:19.157 "base_bdevs_list": [ 00:11:19.157 { 00:11:19.157 "name": "spare", 00:11:19.157 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:19.157 "is_configured": true, 00:11:19.157 "data_offset": 2048, 00:11:19.157 "data_size": 63488 00:11:19.157 }, 00:11:19.157 { 00:11:19.157 "name": "BaseBdev2", 00:11:19.157 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:19.157 "is_configured": true, 00:11:19.157 "data_offset": 2048, 00:11:19.157 "data_size": 63488 00:11:19.157 } 00:11:19.157 ] 00:11:19.157 }' 00:11:19.157 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.157 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:19.157 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:19.157 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:19.157 14:10:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:19.724 [2024-09-30 14:10:24.182500] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:19.724 [2024-09-30 14:10:24.182583] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:19.724 [2024-09-30 14:10:24.182683] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.990 "name": "raid_bdev1", 00:11:19.990 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:19.990 "strip_size_kb": 0, 00:11:19.990 "state": "online", 00:11:19.990 "raid_level": "raid1", 00:11:19.990 "superblock": true, 00:11:19.990 "num_base_bdevs": 2, 00:11:19.990 "num_base_bdevs_discovered": 2, 00:11:19.990 "num_base_bdevs_operational": 2, 00:11:19.990 "base_bdevs_list": [ 00:11:19.990 { 00:11:19.990 "name": "spare", 00:11:19.990 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:19.990 "is_configured": true, 00:11:19.990 "data_offset": 2048, 00:11:19.990 "data_size": 63488 00:11:19.990 }, 00:11:19.990 { 00:11:19.990 "name": "BaseBdev2", 00:11:19.990 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:19.990 "is_configured": true, 00:11:19.990 "data_offset": 2048, 00:11:19.990 "data_size": 63488 00:11:19.990 } 00:11:19.990 ] 00:11:19.990 }' 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:19.990 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:20.251 "name": "raid_bdev1", 00:11:20.251 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:20.251 "strip_size_kb": 0, 00:11:20.251 "state": "online", 00:11:20.251 "raid_level": "raid1", 00:11:20.251 "superblock": true, 00:11:20.251 "num_base_bdevs": 2, 00:11:20.251 "num_base_bdevs_discovered": 2, 00:11:20.251 "num_base_bdevs_operational": 2, 00:11:20.251 "base_bdevs_list": [ 00:11:20.251 { 00:11:20.251 "name": "spare", 00:11:20.251 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:20.251 "is_configured": true, 00:11:20.251 "data_offset": 2048, 00:11:20.251 "data_size": 63488 00:11:20.251 }, 00:11:20.251 { 00:11:20.251 "name": "BaseBdev2", 00:11:20.251 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:20.251 "is_configured": true, 00:11:20.251 "data_offset": 2048, 00:11:20.251 "data_size": 63488 00:11:20.251 } 00:11:20.251 ] 00:11:20.251 }' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.251 "name": "raid_bdev1", 00:11:20.251 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:20.251 "strip_size_kb": 0, 00:11:20.251 "state": "online", 00:11:20.251 "raid_level": "raid1", 00:11:20.251 "superblock": true, 00:11:20.251 "num_base_bdevs": 2, 00:11:20.251 "num_base_bdevs_discovered": 2, 00:11:20.251 "num_base_bdevs_operational": 2, 00:11:20.251 "base_bdevs_list": [ 00:11:20.251 { 00:11:20.251 "name": "spare", 00:11:20.251 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:20.251 "is_configured": true, 00:11:20.251 "data_offset": 2048, 00:11:20.251 "data_size": 63488 00:11:20.251 }, 00:11:20.251 { 00:11:20.251 "name": "BaseBdev2", 00:11:20.251 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:20.251 "is_configured": true, 00:11:20.251 "data_offset": 2048, 00:11:20.251 "data_size": 63488 00:11:20.251 } 00:11:20.251 ] 00:11:20.251 }' 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.251 14:10:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.820 [2024-09-30 14:10:25.173253] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:20.820 [2024-09-30 14:10:25.173289] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:20.820 [2024-09-30 14:10:25.173387] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:20.820 [2024-09-30 14:10:25.173466] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:20.820 [2024-09-30 14:10:25.173494] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:20.820 /dev/nbd0 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:20.820 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:21.086 1+0 records in 00:11:21.086 1+0 records out 00:11:21.086 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000487799 s, 8.4 MB/s 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:21.086 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:21.086 /dev/nbd1 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:21.361 1+0 records in 00:11:21.361 1+0 records out 00:11:21.361 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00028672 s, 14.3 MB/s 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:21.361 14:10:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.628 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.887 [2024-09-30 14:10:26.290988] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:21.887 [2024-09-30 14:10:26.291045] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:21.887 [2024-09-30 14:10:26.291070] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:21.887 [2024-09-30 14:10:26.291080] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:21.887 [2024-09-30 14:10:26.293264] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:21.887 [2024-09-30 14:10:26.293301] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:21.887 [2024-09-30 14:10:26.293379] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:21.887 [2024-09-30 14:10:26.293415] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:21.887 [2024-09-30 14:10:26.293556] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:21.887 spare 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.887 [2024-09-30 14:10:26.393463] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:11:21.887 [2024-09-30 14:10:26.393507] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:21.887 [2024-09-30 14:10:26.393804] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:11:21.887 [2024-09-30 14:10:26.393988] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:11:21.887 [2024-09-30 14:10:26.394005] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:11:21.887 [2024-09-30 14:10:26.394151] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.887 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.887 "name": "raid_bdev1", 00:11:21.887 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:21.887 "strip_size_kb": 0, 00:11:21.887 "state": "online", 00:11:21.887 "raid_level": "raid1", 00:11:21.887 "superblock": true, 00:11:21.887 "num_base_bdevs": 2, 00:11:21.887 "num_base_bdevs_discovered": 2, 00:11:21.887 "num_base_bdevs_operational": 2, 00:11:21.887 "base_bdevs_list": [ 00:11:21.887 { 00:11:21.887 "name": "spare", 00:11:21.887 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:21.887 "is_configured": true, 00:11:21.887 "data_offset": 2048, 00:11:21.887 "data_size": 63488 00:11:21.887 }, 00:11:21.887 { 00:11:21.887 "name": "BaseBdev2", 00:11:21.887 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:21.887 "is_configured": true, 00:11:21.887 "data_offset": 2048, 00:11:21.887 "data_size": 63488 00:11:21.887 } 00:11:21.887 ] 00:11:21.887 }' 00:11:21.888 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.888 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:22.457 "name": "raid_bdev1", 00:11:22.457 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:22.457 "strip_size_kb": 0, 00:11:22.457 "state": "online", 00:11:22.457 "raid_level": "raid1", 00:11:22.457 "superblock": true, 00:11:22.457 "num_base_bdevs": 2, 00:11:22.457 "num_base_bdevs_discovered": 2, 00:11:22.457 "num_base_bdevs_operational": 2, 00:11:22.457 "base_bdevs_list": [ 00:11:22.457 { 00:11:22.457 "name": "spare", 00:11:22.457 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:22.457 "is_configured": true, 00:11:22.457 "data_offset": 2048, 00:11:22.457 "data_size": 63488 00:11:22.457 }, 00:11:22.457 { 00:11:22.457 "name": "BaseBdev2", 00:11:22.457 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:22.457 "is_configured": true, 00:11:22.457 "data_offset": 2048, 00:11:22.457 "data_size": 63488 00:11:22.457 } 00:11:22.457 ] 00:11:22.457 }' 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.457 14:10:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.457 [2024-09-30 14:10:27.029758] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.457 "name": "raid_bdev1", 00:11:22.457 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:22.457 "strip_size_kb": 0, 00:11:22.457 "state": "online", 00:11:22.457 "raid_level": "raid1", 00:11:22.457 "superblock": true, 00:11:22.457 "num_base_bdevs": 2, 00:11:22.457 "num_base_bdevs_discovered": 1, 00:11:22.457 "num_base_bdevs_operational": 1, 00:11:22.457 "base_bdevs_list": [ 00:11:22.457 { 00:11:22.457 "name": null, 00:11:22.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.457 "is_configured": false, 00:11:22.457 "data_offset": 0, 00:11:22.457 "data_size": 63488 00:11:22.457 }, 00:11:22.457 { 00:11:22.457 "name": "BaseBdev2", 00:11:22.457 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:22.457 "is_configured": true, 00:11:22.457 "data_offset": 2048, 00:11:22.457 "data_size": 63488 00:11:22.457 } 00:11:22.457 ] 00:11:22.457 }' 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.457 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.025 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:23.025 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.025 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.025 [2024-09-30 14:10:27.476973] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:23.025 [2024-09-30 14:10:27.477150] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:23.025 [2024-09-30 14:10:27.477168] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:23.025 [2024-09-30 14:10:27.477209] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:23.025 [2024-09-30 14:10:27.481140] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:11:23.025 14:10:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.025 14:10:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:23.025 [2024-09-30 14:10:27.483025] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:23.963 "name": "raid_bdev1", 00:11:23.963 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:23.963 "strip_size_kb": 0, 00:11:23.963 "state": "online", 00:11:23.963 "raid_level": "raid1", 00:11:23.963 "superblock": true, 00:11:23.963 "num_base_bdevs": 2, 00:11:23.963 "num_base_bdevs_discovered": 2, 00:11:23.963 "num_base_bdevs_operational": 2, 00:11:23.963 "process": { 00:11:23.963 "type": "rebuild", 00:11:23.963 "target": "spare", 00:11:23.963 "progress": { 00:11:23.963 "blocks": 20480, 00:11:23.963 "percent": 32 00:11:23.963 } 00:11:23.963 }, 00:11:23.963 "base_bdevs_list": [ 00:11:23.963 { 00:11:23.963 "name": "spare", 00:11:23.963 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:23.963 "is_configured": true, 00:11:23.963 "data_offset": 2048, 00:11:23.963 "data_size": 63488 00:11:23.963 }, 00:11:23.963 { 00:11:23.963 "name": "BaseBdev2", 00:11:23.963 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:23.963 "is_configured": true, 00:11:23.963 "data_offset": 2048, 00:11:23.963 "data_size": 63488 00:11:23.963 } 00:11:23.963 ] 00:11:23.963 }' 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:23.963 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.222 [2024-09-30 14:10:28.647903] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:24.222 [2024-09-30 14:10:28.687107] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:24.222 [2024-09-30 14:10:28.687177] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:24.222 [2024-09-30 14:10:28.687194] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:24.222 [2024-09-30 14:10:28.687205] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.222 "name": "raid_bdev1", 00:11:24.222 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:24.222 "strip_size_kb": 0, 00:11:24.222 "state": "online", 00:11:24.222 "raid_level": "raid1", 00:11:24.222 "superblock": true, 00:11:24.222 "num_base_bdevs": 2, 00:11:24.222 "num_base_bdevs_discovered": 1, 00:11:24.222 "num_base_bdevs_operational": 1, 00:11:24.222 "base_bdevs_list": [ 00:11:24.222 { 00:11:24.222 "name": null, 00:11:24.222 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.222 "is_configured": false, 00:11:24.222 "data_offset": 0, 00:11:24.222 "data_size": 63488 00:11:24.222 }, 00:11:24.222 { 00:11:24.222 "name": "BaseBdev2", 00:11:24.222 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:24.222 "is_configured": true, 00:11:24.222 "data_offset": 2048, 00:11:24.222 "data_size": 63488 00:11:24.222 } 00:11:24.222 ] 00:11:24.222 }' 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.222 14:10:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.790 14:10:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:24.790 14:10:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.790 14:10:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.790 [2024-09-30 14:10:29.202638] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:24.790 [2024-09-30 14:10:29.202711] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.790 [2024-09-30 14:10:29.202734] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:24.790 [2024-09-30 14:10:29.202749] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.790 [2024-09-30 14:10:29.203219] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.790 [2024-09-30 14:10:29.203248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:24.790 [2024-09-30 14:10:29.203339] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:24.790 [2024-09-30 14:10:29.203365] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:24.790 [2024-09-30 14:10:29.203379] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:24.790 [2024-09-30 14:10:29.203410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:24.790 [2024-09-30 14:10:29.207315] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1d50 00:11:24.790 spare 00:11:24.790 14:10:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.790 14:10:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:24.790 [2024-09-30 14:10:29.209313] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:25.727 "name": "raid_bdev1", 00:11:25.727 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:25.727 "strip_size_kb": 0, 00:11:25.727 "state": "online", 00:11:25.727 "raid_level": "raid1", 00:11:25.727 "superblock": true, 00:11:25.727 "num_base_bdevs": 2, 00:11:25.727 "num_base_bdevs_discovered": 2, 00:11:25.727 "num_base_bdevs_operational": 2, 00:11:25.727 "process": { 00:11:25.727 "type": "rebuild", 00:11:25.727 "target": "spare", 00:11:25.727 "progress": { 00:11:25.727 "blocks": 20480, 00:11:25.727 "percent": 32 00:11:25.727 } 00:11:25.727 }, 00:11:25.727 "base_bdevs_list": [ 00:11:25.727 { 00:11:25.727 "name": "spare", 00:11:25.727 "uuid": "2db2b730-eb28-59a3-97f9-1f7ef0e7c85a", 00:11:25.727 "is_configured": true, 00:11:25.727 "data_offset": 2048, 00:11:25.727 "data_size": 63488 00:11:25.727 }, 00:11:25.727 { 00:11:25.727 "name": "BaseBdev2", 00:11:25.727 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:25.727 "is_configured": true, 00:11:25.727 "data_offset": 2048, 00:11:25.727 "data_size": 63488 00:11:25.727 } 00:11:25.727 ] 00:11:25.727 }' 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.727 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.727 [2024-09-30 14:10:30.346023] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:25.987 [2024-09-30 14:10:30.413834] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:25.987 [2024-09-30 14:10:30.413905] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.987 [2024-09-30 14:10:30.413927] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:25.987 [2024-09-30 14:10:30.413936] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.987 "name": "raid_bdev1", 00:11:25.987 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:25.987 "strip_size_kb": 0, 00:11:25.987 "state": "online", 00:11:25.987 "raid_level": "raid1", 00:11:25.987 "superblock": true, 00:11:25.987 "num_base_bdevs": 2, 00:11:25.987 "num_base_bdevs_discovered": 1, 00:11:25.987 "num_base_bdevs_operational": 1, 00:11:25.987 "base_bdevs_list": [ 00:11:25.987 { 00:11:25.987 "name": null, 00:11:25.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.987 "is_configured": false, 00:11:25.987 "data_offset": 0, 00:11:25.987 "data_size": 63488 00:11:25.987 }, 00:11:25.987 { 00:11:25.987 "name": "BaseBdev2", 00:11:25.987 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:25.987 "is_configured": true, 00:11:25.987 "data_offset": 2048, 00:11:25.987 "data_size": 63488 00:11:25.987 } 00:11:25.987 ] 00:11:25.987 }' 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.987 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.246 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.506 14:10:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.506 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:26.506 "name": "raid_bdev1", 00:11:26.506 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:26.506 "strip_size_kb": 0, 00:11:26.506 "state": "online", 00:11:26.506 "raid_level": "raid1", 00:11:26.506 "superblock": true, 00:11:26.506 "num_base_bdevs": 2, 00:11:26.506 "num_base_bdevs_discovered": 1, 00:11:26.506 "num_base_bdevs_operational": 1, 00:11:26.506 "base_bdevs_list": [ 00:11:26.506 { 00:11:26.506 "name": null, 00:11:26.506 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.506 "is_configured": false, 00:11:26.506 "data_offset": 0, 00:11:26.506 "data_size": 63488 00:11:26.506 }, 00:11:26.506 { 00:11:26.506 "name": "BaseBdev2", 00:11:26.506 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:26.506 "is_configured": true, 00:11:26.506 "data_offset": 2048, 00:11:26.506 "data_size": 63488 00:11:26.506 } 00:11:26.506 ] 00:11:26.506 }' 00:11:26.506 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:26.506 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:26.506 14:10:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.506 [2024-09-30 14:10:31.045301] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:26.506 [2024-09-30 14:10:31.045360] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:26.506 [2024-09-30 14:10:31.045385] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:26.506 [2024-09-30 14:10:31.045396] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:26.506 [2024-09-30 14:10:31.045862] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:26.506 [2024-09-30 14:10:31.045891] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:26.506 [2024-09-30 14:10:31.045982] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:26.506 [2024-09-30 14:10:31.046021] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:26.506 [2024-09-30 14:10:31.046032] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:26.506 [2024-09-30 14:10:31.046044] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:26.506 BaseBdev1 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.506 14:10:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:27.584 "name": "raid_bdev1", 00:11:27.584 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:27.584 "strip_size_kb": 0, 00:11:27.584 "state": "online", 00:11:27.584 "raid_level": "raid1", 00:11:27.584 "superblock": true, 00:11:27.584 "num_base_bdevs": 2, 00:11:27.584 "num_base_bdevs_discovered": 1, 00:11:27.584 "num_base_bdevs_operational": 1, 00:11:27.584 "base_bdevs_list": [ 00:11:27.584 { 00:11:27.584 "name": null, 00:11:27.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:27.584 "is_configured": false, 00:11:27.584 "data_offset": 0, 00:11:27.584 "data_size": 63488 00:11:27.584 }, 00:11:27.584 { 00:11:27.584 "name": "BaseBdev2", 00:11:27.584 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:27.584 "is_configured": true, 00:11:27.584 "data_offset": 2048, 00:11:27.584 "data_size": 63488 00:11:27.584 } 00:11:27.584 ] 00:11:27.584 }' 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:27.584 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.862 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:27.862 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:27.862 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:27.862 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:27.862 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:27.863 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.863 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.863 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.863 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:28.121 "name": "raid_bdev1", 00:11:28.121 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:28.121 "strip_size_kb": 0, 00:11:28.121 "state": "online", 00:11:28.121 "raid_level": "raid1", 00:11:28.121 "superblock": true, 00:11:28.121 "num_base_bdevs": 2, 00:11:28.121 "num_base_bdevs_discovered": 1, 00:11:28.121 "num_base_bdevs_operational": 1, 00:11:28.121 "base_bdevs_list": [ 00:11:28.121 { 00:11:28.121 "name": null, 00:11:28.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.121 "is_configured": false, 00:11:28.121 "data_offset": 0, 00:11:28.121 "data_size": 63488 00:11:28.121 }, 00:11:28.121 { 00:11:28.121 "name": "BaseBdev2", 00:11:28.121 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:28.121 "is_configured": true, 00:11:28.121 "data_offset": 2048, 00:11:28.121 "data_size": 63488 00:11:28.121 } 00:11:28.121 ] 00:11:28.121 }' 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.121 [2024-09-30 14:10:32.646992] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:28.121 [2024-09-30 14:10:32.647174] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:28.121 [2024-09-30 14:10:32.647201] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:28.121 request: 00:11:28.121 { 00:11:28.121 "base_bdev": "BaseBdev1", 00:11:28.121 "raid_bdev": "raid_bdev1", 00:11:28.121 "method": "bdev_raid_add_base_bdev", 00:11:28.121 "req_id": 1 00:11:28.121 } 00:11:28.121 Got JSON-RPC error response 00:11:28.121 response: 00:11:28.121 { 00:11:28.121 "code": -22, 00:11:28.121 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:28.121 } 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:28.121 14:10:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.059 "name": "raid_bdev1", 00:11:29.059 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:29.059 "strip_size_kb": 0, 00:11:29.059 "state": "online", 00:11:29.059 "raid_level": "raid1", 00:11:29.059 "superblock": true, 00:11:29.059 "num_base_bdevs": 2, 00:11:29.059 "num_base_bdevs_discovered": 1, 00:11:29.059 "num_base_bdevs_operational": 1, 00:11:29.059 "base_bdevs_list": [ 00:11:29.059 { 00:11:29.059 "name": null, 00:11:29.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.059 "is_configured": false, 00:11:29.059 "data_offset": 0, 00:11:29.059 "data_size": 63488 00:11:29.059 }, 00:11:29.059 { 00:11:29.059 "name": "BaseBdev2", 00:11:29.059 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:29.059 "is_configured": true, 00:11:29.059 "data_offset": 2048, 00:11:29.059 "data_size": 63488 00:11:29.059 } 00:11:29.059 ] 00:11:29.059 }' 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.059 14:10:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:29.629 "name": "raid_bdev1", 00:11:29.629 "uuid": "8da5ce38-0448-478e-8616-16bb2a2e8aea", 00:11:29.629 "strip_size_kb": 0, 00:11:29.629 "state": "online", 00:11:29.629 "raid_level": "raid1", 00:11:29.629 "superblock": true, 00:11:29.629 "num_base_bdevs": 2, 00:11:29.629 "num_base_bdevs_discovered": 1, 00:11:29.629 "num_base_bdevs_operational": 1, 00:11:29.629 "base_bdevs_list": [ 00:11:29.629 { 00:11:29.629 "name": null, 00:11:29.629 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.629 "is_configured": false, 00:11:29.629 "data_offset": 0, 00:11:29.629 "data_size": 63488 00:11:29.629 }, 00:11:29.629 { 00:11:29.629 "name": "BaseBdev2", 00:11:29.629 "uuid": "b40c410d-4bdd-5748-a72c-f7ad89909a99", 00:11:29.629 "is_configured": true, 00:11:29.629 "data_offset": 2048, 00:11:29.629 "data_size": 63488 00:11:29.629 } 00:11:29.629 ] 00:11:29.629 }' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 87445 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 87445 ']' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 87445 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:29.629 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87445 00:11:29.888 killing process with pid 87445 00:11:29.888 Received shutdown signal, test time was about 60.000000 seconds 00:11:29.888 00:11:29.888 Latency(us) 00:11:29.888 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:29.888 =================================================================================================================== 00:11:29.888 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:29.888 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:29.888 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:29.888 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87445' 00:11:29.888 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 87445 00:11:29.888 [2024-09-30 14:10:34.308096] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:29.888 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 87445 00:11:29.888 [2024-09-30 14:10:34.308260] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:29.888 [2024-09-30 14:10:34.308315] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:29.888 [2024-09-30 14:10:34.308327] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:11:29.888 [2024-09-30 14:10:34.339447] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:11:30.147 00:11:30.147 real 0m21.938s 00:11:30.147 user 0m27.093s 00:11:30.147 sys 0m3.818s 00:11:30.147 ************************************ 00:11:30.147 END TEST raid_rebuild_test_sb 00:11:30.147 ************************************ 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.147 14:10:34 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:11:30.147 14:10:34 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:30.147 14:10:34 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:30.147 14:10:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:30.147 ************************************ 00:11:30.147 START TEST raid_rebuild_test_io 00:11:30.147 ************************************ 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false true true 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=88162 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 88162 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 88162 ']' 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:30.147 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:30.147 14:10:34 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.147 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:30.147 Zero copy mechanism will not be used. 00:11:30.147 [2024-09-30 14:10:34.755129] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:11:30.147 [2024-09-30 14:10:34.755268] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88162 ] 00:11:30.406 [2024-09-30 14:10:34.892137] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:11:30.406 [2024-09-30 14:10:34.904431] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:30.406 [2024-09-30 14:10:34.952273] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:30.406 [2024-09-30 14:10:34.996769] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:30.406 [2024-09-30 14:10:34.996815] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 BaseBdev1_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 [2024-09-30 14:10:35.671963] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:31.341 [2024-09-30 14:10:35.672051] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.341 [2024-09-30 14:10:35.672082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:31.341 [2024-09-30 14:10:35.672111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.341 [2024-09-30 14:10:35.674653] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.341 [2024-09-30 14:10:35.674699] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:31.341 BaseBdev1 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 BaseBdev2_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 [2024-09-30 14:10:35.708981] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:31.341 [2024-09-30 14:10:35.709043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.341 [2024-09-30 14:10:35.709065] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:31.341 [2024-09-30 14:10:35.709078] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.341 [2024-09-30 14:10:35.711547] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.341 [2024-09-30 14:10:35.711639] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:31.341 BaseBdev2 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 spare_malloc 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 spare_delay 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 [2024-09-30 14:10:35.750058] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:31.341 [2024-09-30 14:10:35.750124] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.341 [2024-09-30 14:10:35.750146] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:31.341 [2024-09-30 14:10:35.750160] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.341 [2024-09-30 14:10:35.752645] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.341 [2024-09-30 14:10:35.752689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:31.341 spare 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 [2024-09-30 14:10:35.762089] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:31.341 [2024-09-30 14:10:35.764330] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:31.341 [2024-09-30 14:10:35.764430] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:11:31.341 [2024-09-30 14:10:35.764450] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:31.341 [2024-09-30 14:10:35.764754] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:31.341 [2024-09-30 14:10:35.764892] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:11:31.341 [2024-09-30 14:10:35.764911] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:11:31.341 [2024-09-30 14:10:35.765071] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.341 "name": "raid_bdev1", 00:11:31.341 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:31.341 "strip_size_kb": 0, 00:11:31.341 "state": "online", 00:11:31.341 "raid_level": "raid1", 00:11:31.341 "superblock": false, 00:11:31.341 "num_base_bdevs": 2, 00:11:31.341 "num_base_bdevs_discovered": 2, 00:11:31.341 "num_base_bdevs_operational": 2, 00:11:31.341 "base_bdevs_list": [ 00:11:31.341 { 00:11:31.341 "name": "BaseBdev1", 00:11:31.341 "uuid": "e487fe70-ee62-5c08-8b1d-78e1bb327f7a", 00:11:31.341 "is_configured": true, 00:11:31.341 "data_offset": 0, 00:11:31.341 "data_size": 65536 00:11:31.341 }, 00:11:31.341 { 00:11:31.341 "name": "BaseBdev2", 00:11:31.341 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:31.341 "is_configured": true, 00:11:31.341 "data_offset": 0, 00:11:31.341 "data_size": 65536 00:11:31.341 } 00:11:31.341 ] 00:11:31.341 }' 00:11:31.341 14:10:35 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.342 14:10:35 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.600 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:31.600 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:31.600 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.600 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.600 [2024-09-30 14:10:36.237659] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.859 [2024-09-30 14:10:36.337121] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.859 "name": "raid_bdev1", 00:11:31.859 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:31.859 "strip_size_kb": 0, 00:11:31.859 "state": "online", 00:11:31.859 "raid_level": "raid1", 00:11:31.859 "superblock": false, 00:11:31.859 "num_base_bdevs": 2, 00:11:31.859 "num_base_bdevs_discovered": 1, 00:11:31.859 "num_base_bdevs_operational": 1, 00:11:31.859 "base_bdevs_list": [ 00:11:31.859 { 00:11:31.859 "name": null, 00:11:31.859 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:31.859 "is_configured": false, 00:11:31.859 "data_offset": 0, 00:11:31.859 "data_size": 65536 00:11:31.859 }, 00:11:31.859 { 00:11:31.859 "name": "BaseBdev2", 00:11:31.859 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:31.859 "is_configured": true, 00:11:31.859 "data_offset": 0, 00:11:31.859 "data_size": 65536 00:11:31.859 } 00:11:31.859 ] 00:11:31.859 }' 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.859 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.859 [2024-09-30 14:10:36.439010] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:11:31.859 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:31.859 Zero copy mechanism will not be used. 00:11:31.859 Running I/O for 60 seconds... 00:11:32.118 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:32.118 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.118 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:32.376 [2024-09-30 14:10:36.776508] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:32.376 14:10:36 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.376 14:10:36 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:32.376 [2024-09-30 14:10:36.814588] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:11:32.376 [2024-09-30 14:10:36.816886] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:32.376 [2024-09-30 14:10:36.938032] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:32.650 [2024-09-30 14:10:37.173650] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:32.650 [2024-09-30 14:10:37.173946] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:32.909 172.00 IOPS, 516.00 MiB/s [2024-09-30 14:10:37.530020] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:33.167 [2024-09-30 14:10:37.659083] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:33.167 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.168 14:10:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.426 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:33.426 "name": "raid_bdev1", 00:11:33.426 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:33.426 "strip_size_kb": 0, 00:11:33.426 "state": "online", 00:11:33.426 "raid_level": "raid1", 00:11:33.426 "superblock": false, 00:11:33.426 "num_base_bdevs": 2, 00:11:33.426 "num_base_bdevs_discovered": 2, 00:11:33.426 "num_base_bdevs_operational": 2, 00:11:33.426 "process": { 00:11:33.426 "type": "rebuild", 00:11:33.426 "target": "spare", 00:11:33.426 "progress": { 00:11:33.426 "blocks": 10240, 00:11:33.426 "percent": 15 00:11:33.426 } 00:11:33.426 }, 00:11:33.426 "base_bdevs_list": [ 00:11:33.426 { 00:11:33.426 "name": "spare", 00:11:33.426 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:33.426 "is_configured": true, 00:11:33.426 "data_offset": 0, 00:11:33.426 "data_size": 65536 00:11:33.426 }, 00:11:33.426 { 00:11:33.426 "name": "BaseBdev2", 00:11:33.426 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:33.426 "is_configured": true, 00:11:33.426 "data_offset": 0, 00:11:33.426 "data_size": 65536 00:11:33.426 } 00:11:33.426 ] 00:11:33.426 }' 00:11:33.426 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:33.426 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:33.426 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:33.426 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:33.427 14:10:37 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:33.427 14:10:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.427 14:10:37 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.427 [2024-09-30 14:10:37.948966] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:33.685 [2024-09-30 14:10:38.092693] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:33.685 [2024-09-30 14:10:38.101843] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:33.685 [2024-09-30 14:10:38.101937] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:33.685 [2024-09-30 14:10:38.101983] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:33.685 [2024-09-30 14:10:38.108529] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006150 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.685 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.686 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.686 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.686 "name": "raid_bdev1", 00:11:33.686 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:33.686 "strip_size_kb": 0, 00:11:33.686 "state": "online", 00:11:33.686 "raid_level": "raid1", 00:11:33.686 "superblock": false, 00:11:33.686 "num_base_bdevs": 2, 00:11:33.686 "num_base_bdevs_discovered": 1, 00:11:33.686 "num_base_bdevs_operational": 1, 00:11:33.686 "base_bdevs_list": [ 00:11:33.686 { 00:11:33.686 "name": null, 00:11:33.686 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:33.686 "is_configured": false, 00:11:33.686 "data_offset": 0, 00:11:33.686 "data_size": 65536 00:11:33.686 }, 00:11:33.686 { 00:11:33.686 "name": "BaseBdev2", 00:11:33.686 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:33.686 "is_configured": true, 00:11:33.686 "data_offset": 0, 00:11:33.686 "data_size": 65536 00:11:33.686 } 00:11:33.686 ] 00:11:33.686 }' 00:11:33.686 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.686 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.945 162.00 IOPS, 486.00 MiB/s 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.945 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.204 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.204 "name": "raid_bdev1", 00:11:34.204 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:34.204 "strip_size_kb": 0, 00:11:34.204 "state": "online", 00:11:34.204 "raid_level": "raid1", 00:11:34.204 "superblock": false, 00:11:34.204 "num_base_bdevs": 2, 00:11:34.204 "num_base_bdevs_discovered": 1, 00:11:34.204 "num_base_bdevs_operational": 1, 00:11:34.204 "base_bdevs_list": [ 00:11:34.204 { 00:11:34.204 "name": null, 00:11:34.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.204 "is_configured": false, 00:11:34.204 "data_offset": 0, 00:11:34.204 "data_size": 65536 00:11:34.204 }, 00:11:34.204 { 00:11:34.204 "name": "BaseBdev2", 00:11:34.204 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:34.204 "is_configured": true, 00:11:34.204 "data_offset": 0, 00:11:34.204 "data_size": 65536 00:11:34.205 } 00:11:34.205 ] 00:11:34.205 }' 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.205 [2024-09-30 14:10:38.718234] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.205 14:10:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:34.205 [2024-09-30 14:10:38.752231] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:11:34.205 [2024-09-30 14:10:38.754471] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:34.468 [2024-09-30 14:10:38.877674] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:34.468 [2024-09-30 14:10:38.878173] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:34.468 [2024-09-30 14:10:39.002944] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:34.468 [2024-09-30 14:10:39.003171] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:34.728 [2024-09-30 14:10:39.238274] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:34.728 [2024-09-30 14:10:39.238823] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:34.728 [2024-09-30 14:10:39.369123] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:34.728 [2024-09-30 14:10:39.369332] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:35.247 156.00 IOPS, 468.00 MiB/s [2024-09-30 14:10:39.707084] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.247 "name": "raid_bdev1", 00:11:35.247 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:35.247 "strip_size_kb": 0, 00:11:35.247 "state": "online", 00:11:35.247 "raid_level": "raid1", 00:11:35.247 "superblock": false, 00:11:35.247 "num_base_bdevs": 2, 00:11:35.247 "num_base_bdevs_discovered": 2, 00:11:35.247 "num_base_bdevs_operational": 2, 00:11:35.247 "process": { 00:11:35.247 "type": "rebuild", 00:11:35.247 "target": "spare", 00:11:35.247 "progress": { 00:11:35.247 "blocks": 14336, 00:11:35.247 "percent": 21 00:11:35.247 } 00:11:35.247 }, 00:11:35.247 "base_bdevs_list": [ 00:11:35.247 { 00:11:35.247 "name": "spare", 00:11:35.247 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:35.247 "is_configured": true, 00:11:35.247 "data_offset": 0, 00:11:35.247 "data_size": 65536 00:11:35.247 }, 00:11:35.247 { 00:11:35.247 "name": "BaseBdev2", 00:11:35.247 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:35.247 "is_configured": true, 00:11:35.247 "data_offset": 0, 00:11:35.247 "data_size": 65536 00:11:35.247 } 00:11:35.247 ] 00:11:35.247 }' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=320 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.247 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.507 "name": "raid_bdev1", 00:11:35.507 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:35.507 "strip_size_kb": 0, 00:11:35.507 "state": "online", 00:11:35.507 "raid_level": "raid1", 00:11:35.507 "superblock": false, 00:11:35.507 "num_base_bdevs": 2, 00:11:35.507 "num_base_bdevs_discovered": 2, 00:11:35.507 "num_base_bdevs_operational": 2, 00:11:35.507 "process": { 00:11:35.507 "type": "rebuild", 00:11:35.507 "target": "spare", 00:11:35.507 "progress": { 00:11:35.507 "blocks": 14336, 00:11:35.507 "percent": 21 00:11:35.507 } 00:11:35.507 }, 00:11:35.507 "base_bdevs_list": [ 00:11:35.507 { 00:11:35.507 "name": "spare", 00:11:35.507 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:35.507 "is_configured": true, 00:11:35.507 "data_offset": 0, 00:11:35.507 "data_size": 65536 00:11:35.507 }, 00:11:35.507 { 00:11:35.507 "name": "BaseBdev2", 00:11:35.507 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:35.507 "is_configured": true, 00:11:35.507 "data_offset": 0, 00:11:35.507 "data_size": 65536 00:11:35.507 } 00:11:35.507 ] 00:11:35.507 }' 00:11:35.507 [2024-09-30 14:10:39.921155] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:35.507 [2024-09-30 14:10:39.921417] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:35.507 14:10:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:35.766 [2024-09-30 14:10:40.258775] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:36.025 133.50 IOPS, 400.50 MiB/s [2024-09-30 14:10:40.473058] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:36.285 [2024-09-30 14:10:40.686058] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:36.285 [2024-09-30 14:10:40.803751] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.545 "name": "raid_bdev1", 00:11:36.545 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:36.545 "strip_size_kb": 0, 00:11:36.545 "state": "online", 00:11:36.545 "raid_level": "raid1", 00:11:36.545 "superblock": false, 00:11:36.545 "num_base_bdevs": 2, 00:11:36.545 "num_base_bdevs_discovered": 2, 00:11:36.545 "num_base_bdevs_operational": 2, 00:11:36.545 "process": { 00:11:36.545 "type": "rebuild", 00:11:36.545 "target": "spare", 00:11:36.545 "progress": { 00:11:36.545 "blocks": 30720, 00:11:36.545 "percent": 46 00:11:36.545 } 00:11:36.545 }, 00:11:36.545 "base_bdevs_list": [ 00:11:36.545 { 00:11:36.545 "name": "spare", 00:11:36.545 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:36.545 "is_configured": true, 00:11:36.545 "data_offset": 0, 00:11:36.545 "data_size": 65536 00:11:36.545 }, 00:11:36.545 { 00:11:36.545 "name": "BaseBdev2", 00:11:36.545 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:36.545 "is_configured": true, 00:11:36.545 "data_offset": 0, 00:11:36.545 "data_size": 65536 00:11:36.545 } 00:11:36.545 ] 00:11:36.545 }' 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.545 [2024-09-30 14:10:41.117749] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:36.545 [2024-09-30 14:10:41.118119] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:36.545 14:10:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:37.064 116.60 IOPS, 349.80 MiB/s [2024-09-30 14:10:41.564652] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:37.632 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:37.632 "name": "raid_bdev1", 00:11:37.632 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:37.632 "strip_size_kb": 0, 00:11:37.632 "state": "online", 00:11:37.632 "raid_level": "raid1", 00:11:37.632 "superblock": false, 00:11:37.632 "num_base_bdevs": 2, 00:11:37.632 "num_base_bdevs_discovered": 2, 00:11:37.632 "num_base_bdevs_operational": 2, 00:11:37.632 "process": { 00:11:37.632 "type": "rebuild", 00:11:37.632 "target": "spare", 00:11:37.632 "progress": { 00:11:37.632 "blocks": 49152, 00:11:37.632 "percent": 75 00:11:37.632 } 00:11:37.632 }, 00:11:37.632 "base_bdevs_list": [ 00:11:37.632 { 00:11:37.632 "name": "spare", 00:11:37.632 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:37.632 "is_configured": true, 00:11:37.632 "data_offset": 0, 00:11:37.632 "data_size": 65536 00:11:37.632 }, 00:11:37.632 { 00:11:37.632 "name": "BaseBdev2", 00:11:37.633 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:37.633 "is_configured": true, 00:11:37.633 "data_offset": 0, 00:11:37.633 "data_size": 65536 00:11:37.633 } 00:11:37.633 ] 00:11:37.633 }' 00:11:37.633 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.633 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:37.633 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.633 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:37.633 14:10:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:38.459 104.83 IOPS, 314.50 MiB/s [2024-09-30 14:10:42.947160] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:38.459 [2024-09-30 14:10:43.052472] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:38.459 [2024-09-30 14:10:43.054710] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.718 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.719 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.719 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.719 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:38.719 "name": "raid_bdev1", 00:11:38.719 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:38.719 "strip_size_kb": 0, 00:11:38.719 "state": "online", 00:11:38.719 "raid_level": "raid1", 00:11:38.719 "superblock": false, 00:11:38.719 "num_base_bdevs": 2, 00:11:38.719 "num_base_bdevs_discovered": 2, 00:11:38.719 "num_base_bdevs_operational": 2, 00:11:38.719 "base_bdevs_list": [ 00:11:38.719 { 00:11:38.719 "name": "spare", 00:11:38.719 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:38.719 "is_configured": true, 00:11:38.719 "data_offset": 0, 00:11:38.719 "data_size": 65536 00:11:38.719 }, 00:11:38.719 { 00:11:38.719 "name": "BaseBdev2", 00:11:38.719 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:38.719 "is_configured": true, 00:11:38.719 "data_offset": 0, 00:11:38.719 "data_size": 65536 00:11:38.719 } 00:11:38.719 ] 00:11:38.719 }' 00:11:38.719 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.978 97.57 IOPS, 292.71 MiB/s 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:38.978 "name": "raid_bdev1", 00:11:38.978 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:38.978 "strip_size_kb": 0, 00:11:38.978 "state": "online", 00:11:38.978 "raid_level": "raid1", 00:11:38.978 "superblock": false, 00:11:38.978 "num_base_bdevs": 2, 00:11:38.978 "num_base_bdevs_discovered": 2, 00:11:38.978 "num_base_bdevs_operational": 2, 00:11:38.978 "base_bdevs_list": [ 00:11:38.978 { 00:11:38.978 "name": "spare", 00:11:38.978 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:38.978 "is_configured": true, 00:11:38.978 "data_offset": 0, 00:11:38.978 "data_size": 65536 00:11:38.978 }, 00:11:38.978 { 00:11:38.978 "name": "BaseBdev2", 00:11:38.978 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:38.978 "is_configured": true, 00:11:38.978 "data_offset": 0, 00:11:38.978 "data_size": 65536 00:11:38.978 } 00:11:38.978 ] 00:11:38.978 }' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.978 "name": "raid_bdev1", 00:11:38.978 "uuid": "78d95055-c3ea-411e-ad55-6b4c34a7e861", 00:11:38.978 "strip_size_kb": 0, 00:11:38.978 "state": "online", 00:11:38.978 "raid_level": "raid1", 00:11:38.978 "superblock": false, 00:11:38.978 "num_base_bdevs": 2, 00:11:38.978 "num_base_bdevs_discovered": 2, 00:11:38.978 "num_base_bdevs_operational": 2, 00:11:38.978 "base_bdevs_list": [ 00:11:38.978 { 00:11:38.978 "name": "spare", 00:11:38.978 "uuid": "be15902a-7321-5130-9a9e-b580965c1bb1", 00:11:38.978 "is_configured": true, 00:11:38.978 "data_offset": 0, 00:11:38.978 "data_size": 65536 00:11:38.978 }, 00:11:38.978 { 00:11:38.978 "name": "BaseBdev2", 00:11:38.978 "uuid": "3ac8c31c-723f-593e-9780-7cda44da08f8", 00:11:38.978 "is_configured": true, 00:11:38.978 "data_offset": 0, 00:11:38.978 "data_size": 65536 00:11:38.978 } 00:11:38.978 ] 00:11:38.978 }' 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.978 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.547 14:10:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:39.547 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.547 14:10:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.547 [2024-09-30 14:10:43.943685] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:39.547 [2024-09-30 14:10:43.943721] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:39.547 00:11:39.547 Latency(us) 00:11:39.547 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:39.547 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:39.547 raid_bdev1 : 7.60 92.36 277.08 0.00 0.00 14247.19 279.03 112641.79 00:11:39.547 =================================================================================================================== 00:11:39.547 Total : 92.36 277.08 0.00 0.00 14247.19 279.03 112641.79 00:11:39.547 { 00:11:39.547 "results": [ 00:11:39.547 { 00:11:39.547 "job": "raid_bdev1", 00:11:39.547 "core_mask": "0x1", 00:11:39.547 "workload": "randrw", 00:11:39.547 "percentage": 50, 00:11:39.547 "status": "finished", 00:11:39.547 "queue_depth": 2, 00:11:39.547 "io_size": 3145728, 00:11:39.547 "runtime": 7.600597, 00:11:39.547 "iops": 92.36116583999915, 00:11:39.547 "mibps": 277.0834975199974, 00:11:39.547 "io_failed": 0, 00:11:39.547 "io_timeout": 0, 00:11:39.547 "avg_latency_us": 14247.185558417, 00:11:39.547 "min_latency_us": 279.0288209606987, 00:11:39.547 "max_latency_us": 112641.78864628822 00:11:39.547 } 00:11:39.547 ], 00:11:39.547 "core_count": 1 00:11:39.547 } 00:11:39.547 [2024-09-30 14:10:44.030714] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:39.547 [2024-09-30 14:10:44.030753] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:39.547 [2024-09-30 14:10:44.030829] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:39.547 [2024-09-30 14:10:44.030841] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:39.547 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:39.807 /dev/nbd0 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:39.807 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:39.807 1+0 records in 00:11:39.808 1+0 records out 00:11:39.808 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000621157 s, 6.6 MB/s 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:39.808 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:40.068 /dev/nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:40.068 1+0 records in 00:11:40.068 1+0 records out 00:11:40.068 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000211944 s, 19.3 MB/s 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:40.068 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:40.328 14:10:44 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 88162 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 88162 ']' 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 88162 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88162 00:11:40.588 killing process with pid 88162 00:11:40.588 Received shutdown signal, test time was about 8.693109 seconds 00:11:40.588 00:11:40.588 Latency(us) 00:11:40.588 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:40.588 =================================================================================================================== 00:11:40.588 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88162' 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 88162 00:11:40.588 [2024-09-30 14:10:45.117859] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:40.588 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 88162 00:11:40.588 [2024-09-30 14:10:45.144298] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:40.848 00:11:40.848 real 0m10.726s 00:11:40.848 user 0m13.856s 00:11:40.848 sys 0m1.423s 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:40.848 ************************************ 00:11:40.848 END TEST raid_rebuild_test_io 00:11:40.848 ************************************ 00:11:40.848 14:10:45 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:11:40.848 14:10:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:40.848 14:10:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:40.848 14:10:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:40.848 ************************************ 00:11:40.848 START TEST raid_rebuild_test_sb_io 00:11:40.848 ************************************ 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true true true 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=88521 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 88521 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 88521 ']' 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:40.848 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:40.848 14:10:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.108 [2024-09-30 14:10:45.567877] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:11:41.108 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:41.108 Zero copy mechanism will not be used. 00:11:41.108 [2024-09-30 14:10:45.568075] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88521 ] 00:11:41.108 [2024-09-30 14:10:45.704886] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:11:41.108 [2024-09-30 14:10:45.731304] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:41.368 [2024-09-30 14:10:45.778212] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:41.368 [2024-09-30 14:10:45.821216] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.368 [2024-09-30 14:10:45.821254] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 BaseBdev1_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 [2024-09-30 14:10:46.399803] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:41.937 [2024-09-30 14:10:46.399867] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.937 [2024-09-30 14:10:46.399892] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:41.937 [2024-09-30 14:10:46.399908] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.937 [2024-09-30 14:10:46.402039] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.937 [2024-09-30 14:10:46.402077] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:41.937 BaseBdev1 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 BaseBdev2_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 [2024-09-30 14:10:46.442371] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:41.937 [2024-09-30 14:10:46.442564] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.937 [2024-09-30 14:10:46.442624] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:41.937 [2024-09-30 14:10:46.442665] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.937 [2024-09-30 14:10:46.447140] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.937 [2024-09-30 14:10:46.447213] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:41.937 BaseBdev2 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 spare_malloc 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 spare_delay 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 [2024-09-30 14:10:46.485221] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:41.937 [2024-09-30 14:10:46.485271] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.937 [2024-09-30 14:10:46.485305] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:41.937 [2024-09-30 14:10:46.485316] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.937 [2024-09-30 14:10:46.487393] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.937 [2024-09-30 14:10:46.487469] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:41.937 spare 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.937 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.937 [2024-09-30 14:10:46.497254] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:41.937 [2024-09-30 14:10:46.499031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:41.937 [2024-09-30 14:10:46.499232] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:11:41.937 [2024-09-30 14:10:46.499255] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:41.937 [2024-09-30 14:10:46.499496] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:41.937 [2024-09-30 14:10:46.499619] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:11:41.938 [2024-09-30 14:10:46.499636] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:11:41.938 [2024-09-30 14:10:46.499751] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.938 "name": "raid_bdev1", 00:11:41.938 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:41.938 "strip_size_kb": 0, 00:11:41.938 "state": "online", 00:11:41.938 "raid_level": "raid1", 00:11:41.938 "superblock": true, 00:11:41.938 "num_base_bdevs": 2, 00:11:41.938 "num_base_bdevs_discovered": 2, 00:11:41.938 "num_base_bdevs_operational": 2, 00:11:41.938 "base_bdevs_list": [ 00:11:41.938 { 00:11:41.938 "name": "BaseBdev1", 00:11:41.938 "uuid": "07af2926-57ef-5d45-8a47-b3b9d67df59e", 00:11:41.938 "is_configured": true, 00:11:41.938 "data_offset": 2048, 00:11:41.938 "data_size": 63488 00:11:41.938 }, 00:11:41.938 { 00:11:41.938 "name": "BaseBdev2", 00:11:41.938 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:41.938 "is_configured": true, 00:11:41.938 "data_offset": 2048, 00:11:41.938 "data_size": 63488 00:11:41.938 } 00:11:41.938 ] 00:11:41.938 }' 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.938 14:10:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.505 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:42.505 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:42.505 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.505 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.506 [2024-09-30 14:10:47.008645] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.506 [2024-09-30 14:10:47.108171] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.506 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.766 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:42.766 "name": "raid_bdev1", 00:11:42.766 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:42.766 "strip_size_kb": 0, 00:11:42.766 "state": "online", 00:11:42.766 "raid_level": "raid1", 00:11:42.766 "superblock": true, 00:11:42.766 "num_base_bdevs": 2, 00:11:42.766 "num_base_bdevs_discovered": 1, 00:11:42.766 "num_base_bdevs_operational": 1, 00:11:42.766 "base_bdevs_list": [ 00:11:42.766 { 00:11:42.766 "name": null, 00:11:42.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:42.766 "is_configured": false, 00:11:42.766 "data_offset": 0, 00:11:42.766 "data_size": 63488 00:11:42.766 }, 00:11:42.766 { 00:11:42.766 "name": "BaseBdev2", 00:11:42.766 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:42.766 "is_configured": true, 00:11:42.766 "data_offset": 2048, 00:11:42.766 "data_size": 63488 00:11:42.766 } 00:11:42.766 ] 00:11:42.766 }' 00:11:42.766 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:42.766 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.766 [2024-09-30 14:10:47.201942] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:11:42.766 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:42.766 Zero copy mechanism will not be used. 00:11:42.766 Running I/O for 60 seconds... 00:11:43.026 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:43.026 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:43.026 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.026 [2024-09-30 14:10:47.583423] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:43.026 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:43.026 14:10:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:43.026 [2024-09-30 14:10:47.636145] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:11:43.026 [2024-09-30 14:10:47.638105] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:43.285 [2024-09-30 14:10:47.745287] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:43.285 [2024-09-30 14:10:47.745753] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:43.285 [2024-09-30 14:10:47.869153] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:43.285 [2024-09-30 14:10:47.869451] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:43.544 [2024-09-30 14:10:48.198612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:43.544 [2024-09-30 14:10:48.199035] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:43.804 215.00 IOPS, 645.00 MiB/s [2024-09-30 14:10:48.319045] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:44.063 [2024-09-30 14:10:48.556186] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.063 "name": "raid_bdev1", 00:11:44.063 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:44.063 "strip_size_kb": 0, 00:11:44.063 "state": "online", 00:11:44.063 "raid_level": "raid1", 00:11:44.063 "superblock": true, 00:11:44.063 "num_base_bdevs": 2, 00:11:44.063 "num_base_bdevs_discovered": 2, 00:11:44.063 "num_base_bdevs_operational": 2, 00:11:44.063 "process": { 00:11:44.063 "type": "rebuild", 00:11:44.063 "target": "spare", 00:11:44.063 "progress": { 00:11:44.063 "blocks": 14336, 00:11:44.063 "percent": 22 00:11:44.063 } 00:11:44.063 }, 00:11:44.063 "base_bdevs_list": [ 00:11:44.063 { 00:11:44.063 "name": "spare", 00:11:44.063 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:44.063 "is_configured": true, 00:11:44.063 "data_offset": 2048, 00:11:44.063 "data_size": 63488 00:11:44.063 }, 00:11:44.063 { 00:11:44.063 "name": "BaseBdev2", 00:11:44.063 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:44.063 "is_configured": true, 00:11:44.063 "data_offset": 2048, 00:11:44.063 "data_size": 63488 00:11:44.063 } 00:11:44.063 ] 00:11:44.063 }' 00:11:44.063 [2024-09-30 14:10:48.669123] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.063 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.323 [2024-09-30 14:10:48.774514] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:44.323 [2024-09-30 14:10:48.892447] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:44.323 [2024-09-30 14:10:48.899391] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:44.323 [2024-09-30 14:10:48.899428] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:44.323 [2024-09-30 14:10:48.899439] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:44.323 [2024-09-30 14:10:48.915391] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006150 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:44.323 "name": "raid_bdev1", 00:11:44.323 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:44.323 "strip_size_kb": 0, 00:11:44.323 "state": "online", 00:11:44.323 "raid_level": "raid1", 00:11:44.323 "superblock": true, 00:11:44.323 "num_base_bdevs": 2, 00:11:44.323 "num_base_bdevs_discovered": 1, 00:11:44.323 "num_base_bdevs_operational": 1, 00:11:44.323 "base_bdevs_list": [ 00:11:44.323 { 00:11:44.323 "name": null, 00:11:44.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:44.323 "is_configured": false, 00:11:44.323 "data_offset": 0, 00:11:44.323 "data_size": 63488 00:11:44.323 }, 00:11:44.323 { 00:11:44.323 "name": "BaseBdev2", 00:11:44.323 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:44.323 "is_configured": true, 00:11:44.323 "data_offset": 2048, 00:11:44.323 "data_size": 63488 00:11:44.323 } 00:11:44.323 ] 00:11:44.323 }' 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:44.323 14:10:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.849 186.50 IOPS, 559.50 MiB/s 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.849 "name": "raid_bdev1", 00:11:44.849 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:44.849 "strip_size_kb": 0, 00:11:44.849 "state": "online", 00:11:44.849 "raid_level": "raid1", 00:11:44.849 "superblock": true, 00:11:44.849 "num_base_bdevs": 2, 00:11:44.849 "num_base_bdevs_discovered": 1, 00:11:44.849 "num_base_bdevs_operational": 1, 00:11:44.849 "base_bdevs_list": [ 00:11:44.849 { 00:11:44.849 "name": null, 00:11:44.849 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:44.849 "is_configured": false, 00:11:44.849 "data_offset": 0, 00:11:44.849 "data_size": 63488 00:11:44.849 }, 00:11:44.849 { 00:11:44.849 "name": "BaseBdev2", 00:11:44.849 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:44.849 "is_configured": true, 00:11:44.849 "data_offset": 2048, 00:11:44.849 "data_size": 63488 00:11:44.849 } 00:11:44.849 ] 00:11:44.849 }' 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.849 [2024-09-30 14:10:49.438602] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.849 14:10:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:44.849 [2024-09-30 14:10:49.489495] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:11:44.849 [2024-09-30 14:10:49.491337] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:45.109 [2024-09-30 14:10:49.598616] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:45.109 [2024-09-30 14:10:49.599129] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:45.109 [2024-09-30 14:10:49.708086] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:45.109 [2024-09-30 14:10:49.708324] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:45.678 [2024-09-30 14:10:50.051770] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:45.678 [2024-09-30 14:10:50.180280] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:45.678 [2024-09-30 14:10:50.180607] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:45.938 185.00 IOPS, 555.00 MiB/s 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.938 [2024-09-30 14:10:50.519684] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.938 "name": "raid_bdev1", 00:11:45.938 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:45.938 "strip_size_kb": 0, 00:11:45.938 "state": "online", 00:11:45.938 "raid_level": "raid1", 00:11:45.938 "superblock": true, 00:11:45.938 "num_base_bdevs": 2, 00:11:45.938 "num_base_bdevs_discovered": 2, 00:11:45.938 "num_base_bdevs_operational": 2, 00:11:45.938 "process": { 00:11:45.938 "type": "rebuild", 00:11:45.938 "target": "spare", 00:11:45.938 "progress": { 00:11:45.938 "blocks": 12288, 00:11:45.938 "percent": 19 00:11:45.938 } 00:11:45.938 }, 00:11:45.938 "base_bdevs_list": [ 00:11:45.938 { 00:11:45.938 "name": "spare", 00:11:45.938 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:45.938 "is_configured": true, 00:11:45.938 "data_offset": 2048, 00:11:45.938 "data_size": 63488 00:11:45.938 }, 00:11:45.938 { 00:11:45.938 "name": "BaseBdev2", 00:11:45.938 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:45.938 "is_configured": true, 00:11:45.938 "data_offset": 2048, 00:11:45.938 "data_size": 63488 00:11:45.938 } 00:11:45.938 ] 00:11:45.938 }' 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:45.938 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:46.198 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=331 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.198 "name": "raid_bdev1", 00:11:46.198 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:46.198 "strip_size_kb": 0, 00:11:46.198 "state": "online", 00:11:46.198 "raid_level": "raid1", 00:11:46.198 "superblock": true, 00:11:46.198 "num_base_bdevs": 2, 00:11:46.198 "num_base_bdevs_discovered": 2, 00:11:46.198 "num_base_bdevs_operational": 2, 00:11:46.198 "process": { 00:11:46.198 "type": "rebuild", 00:11:46.198 "target": "spare", 00:11:46.198 "progress": { 00:11:46.198 "blocks": 14336, 00:11:46.198 "percent": 22 00:11:46.198 } 00:11:46.198 }, 00:11:46.198 "base_bdevs_list": [ 00:11:46.198 { 00:11:46.198 "name": "spare", 00:11:46.198 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:46.198 "is_configured": true, 00:11:46.198 "data_offset": 2048, 00:11:46.198 "data_size": 63488 00:11:46.198 }, 00:11:46.198 { 00:11:46.198 "name": "BaseBdev2", 00:11:46.198 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:46.198 "is_configured": true, 00:11:46.198 "data_offset": 2048, 00:11:46.198 "data_size": 63488 00:11:46.198 } 00:11:46.198 ] 00:11:46.198 }' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.198 [2024-09-30 14:10:50.738527] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:46.198 [2024-09-30 14:10:50.738808] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:46.198 14:10:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:46.457 [2024-09-30 14:10:51.052558] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:46.716 [2024-09-30 14:10:51.175480] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:46.716 [2024-09-30 14:10:51.175796] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:47.285 156.00 IOPS, 468.00 MiB/s [2024-09-30 14:10:51.636238] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:47.285 [2024-09-30 14:10:51.636564] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:47.285 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:47.286 "name": "raid_bdev1", 00:11:47.286 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:47.286 "strip_size_kb": 0, 00:11:47.286 "state": "online", 00:11:47.286 "raid_level": "raid1", 00:11:47.286 "superblock": true, 00:11:47.286 "num_base_bdevs": 2, 00:11:47.286 "num_base_bdevs_discovered": 2, 00:11:47.286 "num_base_bdevs_operational": 2, 00:11:47.286 "process": { 00:11:47.286 "type": "rebuild", 00:11:47.286 "target": "spare", 00:11:47.286 "progress": { 00:11:47.286 "blocks": 28672, 00:11:47.286 "percent": 45 00:11:47.286 } 00:11:47.286 }, 00:11:47.286 "base_bdevs_list": [ 00:11:47.286 { 00:11:47.286 "name": "spare", 00:11:47.286 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:47.286 "is_configured": true, 00:11:47.286 "data_offset": 2048, 00:11:47.286 "data_size": 63488 00:11:47.286 }, 00:11:47.286 { 00:11:47.286 "name": "BaseBdev2", 00:11:47.286 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:47.286 "is_configured": true, 00:11:47.286 "data_offset": 2048, 00:11:47.286 "data_size": 63488 00:11:47.286 } 00:11:47.286 ] 00:11:47.286 }' 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:47.286 14:10:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:47.545 [2024-09-30 14:10:52.095666] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:47.804 134.80 IOPS, 404.40 MiB/s [2024-09-30 14:10:52.422504] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:48.374 [2024-09-30 14:10:52.754292] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:11:48.374 [2024-09-30 14:10:52.878948] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.374 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:48.374 "name": "raid_bdev1", 00:11:48.374 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:48.374 "strip_size_kb": 0, 00:11:48.374 "state": "online", 00:11:48.374 "raid_level": "raid1", 00:11:48.374 "superblock": true, 00:11:48.374 "num_base_bdevs": 2, 00:11:48.374 "num_base_bdevs_discovered": 2, 00:11:48.374 "num_base_bdevs_operational": 2, 00:11:48.374 "process": { 00:11:48.374 "type": "rebuild", 00:11:48.374 "target": "spare", 00:11:48.374 "progress": { 00:11:48.374 "blocks": 47104, 00:11:48.374 "percent": 74 00:11:48.374 } 00:11:48.374 }, 00:11:48.374 "base_bdevs_list": [ 00:11:48.374 { 00:11:48.374 "name": "spare", 00:11:48.374 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:48.374 "is_configured": true, 00:11:48.374 "data_offset": 2048, 00:11:48.374 "data_size": 63488 00:11:48.374 }, 00:11:48.374 { 00:11:48.374 "name": "BaseBdev2", 00:11:48.374 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:48.374 "is_configured": true, 00:11:48.374 "data_offset": 2048, 00:11:48.374 "data_size": 63488 00:11:48.374 } 00:11:48.374 ] 00:11:48.374 }' 00:11:48.375 14:10:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:48.375 14:10:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:48.375 14:10:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:48.634 14:10:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:48.634 14:10:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:48.634 119.67 IOPS, 359.00 MiB/s [2024-09-30 14:10:53.223325] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:48.894 [2024-09-30 14:10:53.433895] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:48.894 [2024-09-30 14:10:53.434140] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:49.464 [2024-09-30 14:10:53.865092] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.464 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:49.464 "name": "raid_bdev1", 00:11:49.464 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:49.464 "strip_size_kb": 0, 00:11:49.464 "state": "online", 00:11:49.464 "raid_level": "raid1", 00:11:49.464 "superblock": true, 00:11:49.464 "num_base_bdevs": 2, 00:11:49.464 "num_base_bdevs_discovered": 2, 00:11:49.464 "num_base_bdevs_operational": 2, 00:11:49.464 "process": { 00:11:49.464 "type": "rebuild", 00:11:49.464 "target": "spare", 00:11:49.464 "progress": { 00:11:49.465 "blocks": 61440, 00:11:49.465 "percent": 96 00:11:49.465 } 00:11:49.465 }, 00:11:49.465 "base_bdevs_list": [ 00:11:49.465 { 00:11:49.465 "name": "spare", 00:11:49.465 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:49.465 "is_configured": true, 00:11:49.465 "data_offset": 2048, 00:11:49.465 "data_size": 63488 00:11:49.465 }, 00:11:49.465 { 00:11:49.465 "name": "BaseBdev2", 00:11:49.465 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:49.465 "is_configured": true, 00:11:49.465 "data_offset": 2048, 00:11:49.465 "data_size": 63488 00:11:49.465 } 00:11:49.465 ] 00:11:49.465 }' 00:11:49.725 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:49.725 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:49.725 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:49.725 [2024-09-30 14:10:54.186126] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:49.725 107.86 IOPS, 323.57 MiB/s 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:49.725 14:10:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:49.725 [2024-09-30 14:10:54.285988] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:49.725 [2024-09-30 14:10:54.287399] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:50.666 100.25 IOPS, 300.75 MiB/s 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.666 "name": "raid_bdev1", 00:11:50.666 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:50.666 "strip_size_kb": 0, 00:11:50.666 "state": "online", 00:11:50.666 "raid_level": "raid1", 00:11:50.666 "superblock": true, 00:11:50.666 "num_base_bdevs": 2, 00:11:50.666 "num_base_bdevs_discovered": 2, 00:11:50.666 "num_base_bdevs_operational": 2, 00:11:50.666 "base_bdevs_list": [ 00:11:50.666 { 00:11:50.666 "name": "spare", 00:11:50.666 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:50.666 "is_configured": true, 00:11:50.666 "data_offset": 2048, 00:11:50.666 "data_size": 63488 00:11:50.666 }, 00:11:50.666 { 00:11:50.666 "name": "BaseBdev2", 00:11:50.666 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:50.666 "is_configured": true, 00:11:50.666 "data_offset": 2048, 00:11:50.666 "data_size": 63488 00:11:50.666 } 00:11:50.666 ] 00:11:50.666 }' 00:11:50.666 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.935 "name": "raid_bdev1", 00:11:50.935 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:50.935 "strip_size_kb": 0, 00:11:50.935 "state": "online", 00:11:50.935 "raid_level": "raid1", 00:11:50.935 "superblock": true, 00:11:50.935 "num_base_bdevs": 2, 00:11:50.935 "num_base_bdevs_discovered": 2, 00:11:50.935 "num_base_bdevs_operational": 2, 00:11:50.935 "base_bdevs_list": [ 00:11:50.935 { 00:11:50.935 "name": "spare", 00:11:50.935 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:50.935 "is_configured": true, 00:11:50.935 "data_offset": 2048, 00:11:50.935 "data_size": 63488 00:11:50.935 }, 00:11:50.935 { 00:11:50.935 "name": "BaseBdev2", 00:11:50.935 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:50.935 "is_configured": true, 00:11:50.935 "data_offset": 2048, 00:11:50.935 "data_size": 63488 00:11:50.935 } 00:11:50.935 ] 00:11:50.935 }' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.935 "name": "raid_bdev1", 00:11:50.935 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:50.935 "strip_size_kb": 0, 00:11:50.935 "state": "online", 00:11:50.935 "raid_level": "raid1", 00:11:50.935 "superblock": true, 00:11:50.935 "num_base_bdevs": 2, 00:11:50.935 "num_base_bdevs_discovered": 2, 00:11:50.935 "num_base_bdevs_operational": 2, 00:11:50.935 "base_bdevs_list": [ 00:11:50.935 { 00:11:50.935 "name": "spare", 00:11:50.935 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:50.935 "is_configured": true, 00:11:50.935 "data_offset": 2048, 00:11:50.935 "data_size": 63488 00:11:50.935 }, 00:11:50.935 { 00:11:50.935 "name": "BaseBdev2", 00:11:50.935 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:50.935 "is_configured": true, 00:11:50.935 "data_offset": 2048, 00:11:50.935 "data_size": 63488 00:11:50.935 } 00:11:50.935 ] 00:11:50.935 }' 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.935 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.558 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:51.558 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:51.558 14:10:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.558 [2024-09-30 14:10:55.965798] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:51.558 [2024-09-30 14:10:55.965906] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:51.558 00:11:51.558 Latency(us) 00:11:51.558 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.558 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:51.558 raid_bdev1 : 8.82 94.35 283.04 0.00 0.00 14776.91 270.09 107604.96 00:11:51.558 =================================================================================================================== 00:11:51.558 Total : 94.35 283.04 0.00 0.00 14776.91 270.09 107604.96 00:11:51.558 { 00:11:51.558 "results": [ 00:11:51.558 { 00:11:51.558 "job": "raid_bdev1", 00:11:51.558 "core_mask": "0x1", 00:11:51.558 "workload": "randrw", 00:11:51.558 "percentage": 50, 00:11:51.558 "status": "finished", 00:11:51.558 "queue_depth": 2, 00:11:51.558 "io_size": 3145728, 00:11:51.558 "runtime": 8.818691, 00:11:51.558 "iops": 94.34506776572623, 00:11:51.558 "mibps": 283.03520329717867, 00:11:51.558 "io_failed": 0, 00:11:51.558 "io_timeout": 0, 00:11:51.558 "avg_latency_us": 14776.908834397043, 00:11:51.558 "min_latency_us": 270.0855895196507, 00:11:51.558 "max_latency_us": 107604.96069868996 00:11:51.558 } 00:11:51.558 ], 00:11:51.558 "core_count": 1 00:11:51.558 } 00:11:51.558 [2024-09-30 14:10:56.008794] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:51.558 [2024-09-30 14:10:56.008828] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:51.558 [2024-09-30 14:10:56.008918] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:51.558 [2024-09-30 14:10:56.008933] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.558 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:51.846 /dev/nbd0 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:51.846 1+0 records in 00:11:51.846 1+0 records out 00:11:51.846 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000443827 s, 9.2 MB/s 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.846 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:52.120 /dev/nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:52.120 1+0 records in 00:11:52.120 1+0 records out 00:11:52.120 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000525895 s, 7.8 MB/s 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:52.120 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:52.380 14:10:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.640 [2024-09-30 14:10:57.121937] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:52.640 [2024-09-30 14:10:57.121991] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.640 [2024-09-30 14:10:57.122014] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:52.640 [2024-09-30 14:10:57.122023] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.640 [2024-09-30 14:10:57.124103] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.640 [2024-09-30 14:10:57.124183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:52.640 [2024-09-30 14:10:57.124273] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:52.640 [2024-09-30 14:10:57.124313] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:52.640 [2024-09-30 14:10:57.124445] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:52.640 spare 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.640 [2024-09-30 14:10:57.224363] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:11:52.640 [2024-09-30 14:10:57.224390] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:52.640 [2024-09-30 14:10:57.224640] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b1a0 00:11:52.640 [2024-09-30 14:10:57.224767] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:11:52.640 [2024-09-30 14:10:57.224776] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:11:52.640 [2024-09-30 14:10:57.224913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.640 "name": "raid_bdev1", 00:11:52.640 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:52.640 "strip_size_kb": 0, 00:11:52.640 "state": "online", 00:11:52.640 "raid_level": "raid1", 00:11:52.640 "superblock": true, 00:11:52.640 "num_base_bdevs": 2, 00:11:52.640 "num_base_bdevs_discovered": 2, 00:11:52.640 "num_base_bdevs_operational": 2, 00:11:52.640 "base_bdevs_list": [ 00:11:52.640 { 00:11:52.640 "name": "spare", 00:11:52.640 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:52.640 "is_configured": true, 00:11:52.640 "data_offset": 2048, 00:11:52.640 "data_size": 63488 00:11:52.640 }, 00:11:52.640 { 00:11:52.640 "name": "BaseBdev2", 00:11:52.640 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:52.640 "is_configured": true, 00:11:52.640 "data_offset": 2048, 00:11:52.640 "data_size": 63488 00:11:52.640 } 00:11:52.640 ] 00:11:52.640 }' 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.640 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:53.209 "name": "raid_bdev1", 00:11:53.209 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:53.209 "strip_size_kb": 0, 00:11:53.209 "state": "online", 00:11:53.209 "raid_level": "raid1", 00:11:53.209 "superblock": true, 00:11:53.209 "num_base_bdevs": 2, 00:11:53.209 "num_base_bdevs_discovered": 2, 00:11:53.209 "num_base_bdevs_operational": 2, 00:11:53.209 "base_bdevs_list": [ 00:11:53.209 { 00:11:53.209 "name": "spare", 00:11:53.209 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:53.209 "is_configured": true, 00:11:53.209 "data_offset": 2048, 00:11:53.209 "data_size": 63488 00:11:53.209 }, 00:11:53.209 { 00:11:53.209 "name": "BaseBdev2", 00:11:53.209 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:53.209 "is_configured": true, 00:11:53.209 "data_offset": 2048, 00:11:53.209 "data_size": 63488 00:11:53.209 } 00:11:53.209 ] 00:11:53.209 }' 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.209 [2024-09-30 14:10:57.848777] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.209 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.468 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.468 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.468 "name": "raid_bdev1", 00:11:53.468 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:53.468 "strip_size_kb": 0, 00:11:53.468 "state": "online", 00:11:53.468 "raid_level": "raid1", 00:11:53.468 "superblock": true, 00:11:53.468 "num_base_bdevs": 2, 00:11:53.468 "num_base_bdevs_discovered": 1, 00:11:53.468 "num_base_bdevs_operational": 1, 00:11:53.468 "base_bdevs_list": [ 00:11:53.468 { 00:11:53.468 "name": null, 00:11:53.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.468 "is_configured": false, 00:11:53.468 "data_offset": 0, 00:11:53.468 "data_size": 63488 00:11:53.468 }, 00:11:53.468 { 00:11:53.468 "name": "BaseBdev2", 00:11:53.468 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:53.468 "is_configured": true, 00:11:53.468 "data_offset": 2048, 00:11:53.468 "data_size": 63488 00:11:53.468 } 00:11:53.468 ] 00:11:53.468 }' 00:11:53.468 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.468 14:10:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.727 14:10:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:53.727 14:10:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.727 14:10:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.727 [2024-09-30 14:10:58.324098] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:53.727 [2024-09-30 14:10:58.324330] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:53.727 [2024-09-30 14:10:58.324391] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:53.727 [2024-09-30 14:10:58.324462] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:53.727 [2024-09-30 14:10:58.328778] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b270 00:11:53.727 14:10:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.727 14:10:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:53.727 [2024-09-30 14:10:58.330616] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.101 "name": "raid_bdev1", 00:11:55.101 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:55.101 "strip_size_kb": 0, 00:11:55.101 "state": "online", 00:11:55.101 "raid_level": "raid1", 00:11:55.101 "superblock": true, 00:11:55.101 "num_base_bdevs": 2, 00:11:55.101 "num_base_bdevs_discovered": 2, 00:11:55.101 "num_base_bdevs_operational": 2, 00:11:55.101 "process": { 00:11:55.101 "type": "rebuild", 00:11:55.101 "target": "spare", 00:11:55.101 "progress": { 00:11:55.101 "blocks": 20480, 00:11:55.101 "percent": 32 00:11:55.101 } 00:11:55.101 }, 00:11:55.101 "base_bdevs_list": [ 00:11:55.101 { 00:11:55.101 "name": "spare", 00:11:55.101 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:55.101 "is_configured": true, 00:11:55.101 "data_offset": 2048, 00:11:55.101 "data_size": 63488 00:11:55.101 }, 00:11:55.101 { 00:11:55.101 "name": "BaseBdev2", 00:11:55.101 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:55.101 "is_configured": true, 00:11:55.101 "data_offset": 2048, 00:11:55.101 "data_size": 63488 00:11:55.101 } 00:11:55.101 ] 00:11:55.101 }' 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.101 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.102 [2024-09-30 14:10:59.482963] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.102 [2024-09-30 14:10:59.534556] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:55.102 [2024-09-30 14:10:59.534627] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:55.102 [2024-09-30 14:10:59.534641] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.102 [2024-09-30 14:10:59.534650] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.102 "name": "raid_bdev1", 00:11:55.102 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:55.102 "strip_size_kb": 0, 00:11:55.102 "state": "online", 00:11:55.102 "raid_level": "raid1", 00:11:55.102 "superblock": true, 00:11:55.102 "num_base_bdevs": 2, 00:11:55.102 "num_base_bdevs_discovered": 1, 00:11:55.102 "num_base_bdevs_operational": 1, 00:11:55.102 "base_bdevs_list": [ 00:11:55.102 { 00:11:55.102 "name": null, 00:11:55.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.102 "is_configured": false, 00:11:55.102 "data_offset": 0, 00:11:55.102 "data_size": 63488 00:11:55.102 }, 00:11:55.102 { 00:11:55.102 "name": "BaseBdev2", 00:11:55.102 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:55.102 "is_configured": true, 00:11:55.102 "data_offset": 2048, 00:11:55.102 "data_size": 63488 00:11:55.102 } 00:11:55.102 ] 00:11:55.102 }' 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.102 14:10:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.360 14:11:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:55.360 14:11:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.360 14:11:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.360 [2024-09-30 14:11:00.006196] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:55.360 [2024-09-30 14:11:00.006310] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:55.361 [2024-09-30 14:11:00.006348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:55.361 [2024-09-30 14:11:00.006383] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:55.361 [2024-09-30 14:11:00.006840] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:55.361 [2024-09-30 14:11:00.006904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:55.361 [2024-09-30 14:11:00.007045] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:55.361 [2024-09-30 14:11:00.007092] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:55.361 [2024-09-30 14:11:00.007134] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:55.361 [2024-09-30 14:11:00.007227] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:55.361 [2024-09-30 14:11:00.011603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b340 00:11:55.361 spare 00:11:55.361 14:11:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.361 14:11:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:55.361 [2024-09-30 14:11:00.013498] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:56.738 "name": "raid_bdev1", 00:11:56.738 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:56.738 "strip_size_kb": 0, 00:11:56.738 "state": "online", 00:11:56.738 "raid_level": "raid1", 00:11:56.738 "superblock": true, 00:11:56.738 "num_base_bdevs": 2, 00:11:56.738 "num_base_bdevs_discovered": 2, 00:11:56.738 "num_base_bdevs_operational": 2, 00:11:56.738 "process": { 00:11:56.738 "type": "rebuild", 00:11:56.738 "target": "spare", 00:11:56.738 "progress": { 00:11:56.738 "blocks": 20480, 00:11:56.738 "percent": 32 00:11:56.738 } 00:11:56.738 }, 00:11:56.738 "base_bdevs_list": [ 00:11:56.738 { 00:11:56.738 "name": "spare", 00:11:56.738 "uuid": "dd0c4ce1-df3c-5f1a-9445-d7930419f549", 00:11:56.738 "is_configured": true, 00:11:56.738 "data_offset": 2048, 00:11:56.738 "data_size": 63488 00:11:56.738 }, 00:11:56.738 { 00:11:56.738 "name": "BaseBdev2", 00:11:56.738 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:56.738 "is_configured": true, 00:11:56.738 "data_offset": 2048, 00:11:56.738 "data_size": 63488 00:11:56.738 } 00:11:56.738 ] 00:11:56.738 }' 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.738 [2024-09-30 14:11:01.173719] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:56.738 [2024-09-30 14:11:01.217567] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:56.738 [2024-09-30 14:11:01.217679] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:56.738 [2024-09-30 14:11:01.217717] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:56.738 [2024-09-30 14:11:01.217738] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:56.738 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:56.738 "name": "raid_bdev1", 00:11:56.739 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:56.739 "strip_size_kb": 0, 00:11:56.739 "state": "online", 00:11:56.739 "raid_level": "raid1", 00:11:56.739 "superblock": true, 00:11:56.739 "num_base_bdevs": 2, 00:11:56.739 "num_base_bdevs_discovered": 1, 00:11:56.739 "num_base_bdevs_operational": 1, 00:11:56.739 "base_bdevs_list": [ 00:11:56.739 { 00:11:56.739 "name": null, 00:11:56.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.739 "is_configured": false, 00:11:56.739 "data_offset": 0, 00:11:56.739 "data_size": 63488 00:11:56.739 }, 00:11:56.739 { 00:11:56.739 "name": "BaseBdev2", 00:11:56.739 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:56.739 "is_configured": true, 00:11:56.739 "data_offset": 2048, 00:11:56.739 "data_size": 63488 00:11:56.739 } 00:11:56.739 ] 00:11:56.739 }' 00:11:56.739 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:56.739 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.308 "name": "raid_bdev1", 00:11:57.308 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:57.308 "strip_size_kb": 0, 00:11:57.308 "state": "online", 00:11:57.308 "raid_level": "raid1", 00:11:57.308 "superblock": true, 00:11:57.308 "num_base_bdevs": 2, 00:11:57.308 "num_base_bdevs_discovered": 1, 00:11:57.308 "num_base_bdevs_operational": 1, 00:11:57.308 "base_bdevs_list": [ 00:11:57.308 { 00:11:57.308 "name": null, 00:11:57.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.308 "is_configured": false, 00:11:57.308 "data_offset": 0, 00:11:57.308 "data_size": 63488 00:11:57.308 }, 00:11:57.308 { 00:11:57.308 "name": "BaseBdev2", 00:11:57.308 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:57.308 "is_configured": true, 00:11:57.308 "data_offset": 2048, 00:11:57.308 "data_size": 63488 00:11:57.308 } 00:11:57.308 ] 00:11:57.308 }' 00:11:57.308 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.309 [2024-09-30 14:11:01.837184] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:57.309 [2024-09-30 14:11:01.837272] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.309 [2024-09-30 14:11:01.837313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:11:57.309 [2024-09-30 14:11:01.837321] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.309 [2024-09-30 14:11:01.837724] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.309 [2024-09-30 14:11:01.837742] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:57.309 [2024-09-30 14:11:01.837813] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:57.309 [2024-09-30 14:11:01.837828] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:57.309 [2024-09-30 14:11:01.837839] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:57.309 [2024-09-30 14:11:01.837849] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:57.309 BaseBdev1 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.309 14:11:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.247 "name": "raid_bdev1", 00:11:58.247 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:58.247 "strip_size_kb": 0, 00:11:58.247 "state": "online", 00:11:58.247 "raid_level": "raid1", 00:11:58.247 "superblock": true, 00:11:58.247 "num_base_bdevs": 2, 00:11:58.247 "num_base_bdevs_discovered": 1, 00:11:58.247 "num_base_bdevs_operational": 1, 00:11:58.247 "base_bdevs_list": [ 00:11:58.247 { 00:11:58.247 "name": null, 00:11:58.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.247 "is_configured": false, 00:11:58.247 "data_offset": 0, 00:11:58.247 "data_size": 63488 00:11:58.247 }, 00:11:58.247 { 00:11:58.247 "name": "BaseBdev2", 00:11:58.247 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:58.247 "is_configured": true, 00:11:58.247 "data_offset": 2048, 00:11:58.247 "data_size": 63488 00:11:58.247 } 00:11:58.247 ] 00:11:58.247 }' 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.247 14:11:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.815 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:58.815 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.816 "name": "raid_bdev1", 00:11:58.816 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:11:58.816 "strip_size_kb": 0, 00:11:58.816 "state": "online", 00:11:58.816 "raid_level": "raid1", 00:11:58.816 "superblock": true, 00:11:58.816 "num_base_bdevs": 2, 00:11:58.816 "num_base_bdevs_discovered": 1, 00:11:58.816 "num_base_bdevs_operational": 1, 00:11:58.816 "base_bdevs_list": [ 00:11:58.816 { 00:11:58.816 "name": null, 00:11:58.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.816 "is_configured": false, 00:11:58.816 "data_offset": 0, 00:11:58.816 "data_size": 63488 00:11:58.816 }, 00:11:58.816 { 00:11:58.816 "name": "BaseBdev2", 00:11:58.816 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:11:58.816 "is_configured": true, 00:11:58.816 "data_offset": 2048, 00:11:58.816 "data_size": 63488 00:11:58.816 } 00:11:58.816 ] 00:11:58.816 }' 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.816 [2024-09-30 14:11:03.450757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:58.816 [2024-09-30 14:11:03.450960] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:58.816 [2024-09-30 14:11:03.451047] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:58.816 request: 00:11:58.816 { 00:11:58.816 "base_bdev": "BaseBdev1", 00:11:58.816 "raid_bdev": "raid_bdev1", 00:11:58.816 "method": "bdev_raid_add_base_bdev", 00:11:58.816 "req_id": 1 00:11:58.816 } 00:11:58.816 Got JSON-RPC error response 00:11:58.816 response: 00:11:58.816 { 00:11:58.816 "code": -22, 00:11:58.816 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:58.816 } 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:58.816 14:11:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.197 "name": "raid_bdev1", 00:12:00.197 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:12:00.197 "strip_size_kb": 0, 00:12:00.197 "state": "online", 00:12:00.197 "raid_level": "raid1", 00:12:00.197 "superblock": true, 00:12:00.197 "num_base_bdevs": 2, 00:12:00.197 "num_base_bdevs_discovered": 1, 00:12:00.197 "num_base_bdevs_operational": 1, 00:12:00.197 "base_bdevs_list": [ 00:12:00.197 { 00:12:00.197 "name": null, 00:12:00.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.197 "is_configured": false, 00:12:00.197 "data_offset": 0, 00:12:00.197 "data_size": 63488 00:12:00.197 }, 00:12:00.197 { 00:12:00.197 "name": "BaseBdev2", 00:12:00.197 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:12:00.197 "is_configured": true, 00:12:00.197 "data_offset": 2048, 00:12:00.197 "data_size": 63488 00:12:00.197 } 00:12:00.197 ] 00:12:00.197 }' 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.197 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:00.457 "name": "raid_bdev1", 00:12:00.457 "uuid": "79d13c88-e824-4521-92b1-0600d7c51a2b", 00:12:00.457 "strip_size_kb": 0, 00:12:00.457 "state": "online", 00:12:00.457 "raid_level": "raid1", 00:12:00.457 "superblock": true, 00:12:00.457 "num_base_bdevs": 2, 00:12:00.457 "num_base_bdevs_discovered": 1, 00:12:00.457 "num_base_bdevs_operational": 1, 00:12:00.457 "base_bdevs_list": [ 00:12:00.457 { 00:12:00.457 "name": null, 00:12:00.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.457 "is_configured": false, 00:12:00.457 "data_offset": 0, 00:12:00.457 "data_size": 63488 00:12:00.457 }, 00:12:00.457 { 00:12:00.457 "name": "BaseBdev2", 00:12:00.457 "uuid": "9b4ad9af-d76b-53a2-b6f8-750f2e06f0ac", 00:12:00.457 "is_configured": true, 00:12:00.457 "data_offset": 2048, 00:12:00.457 "data_size": 63488 00:12:00.457 } 00:12:00.457 ] 00:12:00.457 }' 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:00.457 14:11:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 88521 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 88521 ']' 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 88521 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88521 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88521' 00:12:00.457 killing process with pid 88521 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 88521 00:12:00.457 Received shutdown signal, test time was about 17.882865 seconds 00:12:00.457 00:12:00.457 Latency(us) 00:12:00.457 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:00.457 =================================================================================================================== 00:12:00.457 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:00.457 [2024-09-30 14:11:05.052385] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:00.457 [2024-09-30 14:11:05.052557] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:00.457 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 88521 00:12:00.457 [2024-09-30 14:11:05.052631] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:00.457 [2024-09-30 14:11:05.052646] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:12:00.457 [2024-09-30 14:11:05.078587] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:00.718 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:00.718 00:12:00.718 real 0m19.854s 00:12:00.718 user 0m26.248s 00:12:00.718 sys 0m2.230s 00:12:00.718 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:00.718 14:11:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.718 ************************************ 00:12:00.718 END TEST raid_rebuild_test_sb_io 00:12:00.718 ************************************ 00:12:00.978 14:11:05 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:00.978 14:11:05 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:12:00.978 14:11:05 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:00.978 14:11:05 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:00.978 14:11:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:00.978 ************************************ 00:12:00.978 START TEST raid_rebuild_test 00:12:00.978 ************************************ 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false false true 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=89220 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 89220 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 89220 ']' 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:00.978 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:00.978 14:11:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:00.978 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:00.978 Zero copy mechanism will not be used. 00:12:00.978 [2024-09-30 14:11:05.494723] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:12:00.978 [2024-09-30 14:11:05.494842] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89220 ] 00:12:00.978 [2024-09-30 14:11:05.624693] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:12:01.237 [2024-09-30 14:11:05.648760] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:01.237 [2024-09-30 14:11:05.692357] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:01.237 [2024-09-30 14:11:05.732819] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:01.237 [2024-09-30 14:11:05.732856] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.805 BaseBdev1_malloc 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.805 [2024-09-30 14:11:06.326075] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:01.805 [2024-09-30 14:11:06.326138] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.805 [2024-09-30 14:11:06.326163] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:01.805 [2024-09-30 14:11:06.326179] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.805 [2024-09-30 14:11:06.328305] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.805 [2024-09-30 14:11:06.328343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:01.805 BaseBdev1 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.805 BaseBdev2_malloc 00:12:01.805 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 [2024-09-30 14:11:06.369978] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:01.806 [2024-09-30 14:11:06.370088] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.806 [2024-09-30 14:11:06.370133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:01.806 [2024-09-30 14:11:06.370158] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.806 [2024-09-30 14:11:06.374957] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.806 [2024-09-30 14:11:06.375048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:01.806 BaseBdev2 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 BaseBdev3_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 [2024-09-30 14:11:06.400965] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:01.806 [2024-09-30 14:11:06.401055] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.806 [2024-09-30 14:11:06.401091] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:01.806 [2024-09-30 14:11:06.401120] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.806 [2024-09-30 14:11:06.403086] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.806 [2024-09-30 14:11:06.403160] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:01.806 BaseBdev3 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 BaseBdev4_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 [2024-09-30 14:11:06.429342] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:01.806 [2024-09-30 14:11:06.429390] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:01.806 [2024-09-30 14:11:06.429423] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:01.806 [2024-09-30 14:11:06.429433] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:01.806 [2024-09-30 14:11:06.431446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:01.806 [2024-09-30 14:11:06.431546] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:01.806 BaseBdev4 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.806 spare_malloc 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.806 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.065 spare_delay 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.065 [2024-09-30 14:11:06.469713] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:02.065 [2024-09-30 14:11:06.469763] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:02.065 [2024-09-30 14:11:06.469781] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:02.065 [2024-09-30 14:11:06.469791] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:02.065 [2024-09-30 14:11:06.471766] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:02.065 [2024-09-30 14:11:06.471802] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:02.065 spare 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.065 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.065 [2024-09-30 14:11:06.481761] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:02.065 [2024-09-30 14:11:06.483490] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:02.065 [2024-09-30 14:11:06.483565] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:02.065 [2024-09-30 14:11:06.483610] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:02.065 [2024-09-30 14:11:06.483679] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:02.065 [2024-09-30 14:11:06.483696] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:02.065 [2024-09-30 14:11:06.483930] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:02.066 [2024-09-30 14:11:06.484061] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:02.066 [2024-09-30 14:11:06.484070] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:02.066 [2024-09-30 14:11:06.484180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.066 "name": "raid_bdev1", 00:12:02.066 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:02.066 "strip_size_kb": 0, 00:12:02.066 "state": "online", 00:12:02.066 "raid_level": "raid1", 00:12:02.066 "superblock": false, 00:12:02.066 "num_base_bdevs": 4, 00:12:02.066 "num_base_bdevs_discovered": 4, 00:12:02.066 "num_base_bdevs_operational": 4, 00:12:02.066 "base_bdevs_list": [ 00:12:02.066 { 00:12:02.066 "name": "BaseBdev1", 00:12:02.066 "uuid": "57e32d8e-f2bb-5e71-899a-41d512cdd2cf", 00:12:02.066 "is_configured": true, 00:12:02.066 "data_offset": 0, 00:12:02.066 "data_size": 65536 00:12:02.066 }, 00:12:02.066 { 00:12:02.066 "name": "BaseBdev2", 00:12:02.066 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:02.066 "is_configured": true, 00:12:02.066 "data_offset": 0, 00:12:02.066 "data_size": 65536 00:12:02.066 }, 00:12:02.066 { 00:12:02.066 "name": "BaseBdev3", 00:12:02.066 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:02.066 "is_configured": true, 00:12:02.066 "data_offset": 0, 00:12:02.066 "data_size": 65536 00:12:02.066 }, 00:12:02.066 { 00:12:02.066 "name": "BaseBdev4", 00:12:02.066 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:02.066 "is_configured": true, 00:12:02.066 "data_offset": 0, 00:12:02.066 "data_size": 65536 00:12:02.066 } 00:12:02.066 ] 00:12:02.066 }' 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.066 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.325 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:02.325 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:02.325 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.325 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.325 [2024-09-30 14:11:06.945217] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:02.325 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.584 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:02.584 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:02.584 14:11:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.584 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.584 14:11:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:02.584 [2024-09-30 14:11:07.200585] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:12:02.584 /dev/nbd0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:02.584 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:02.844 1+0 records in 00:12:02.844 1+0 records out 00:12:02.844 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000298166 s, 13.7 MB/s 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:02.844 14:11:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:08.124 65536+0 records in 00:12:08.124 65536+0 records out 00:12:08.124 33554432 bytes (34 MB, 32 MiB) copied, 4.87005 s, 6.9 MB/s 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:08.124 [2024-09-30 14:11:12.330361] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.124 [2024-09-30 14:11:12.366633] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.124 "name": "raid_bdev1", 00:12:08.124 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:08.124 "strip_size_kb": 0, 00:12:08.124 "state": "online", 00:12:08.124 "raid_level": "raid1", 00:12:08.124 "superblock": false, 00:12:08.124 "num_base_bdevs": 4, 00:12:08.124 "num_base_bdevs_discovered": 3, 00:12:08.124 "num_base_bdevs_operational": 3, 00:12:08.124 "base_bdevs_list": [ 00:12:08.124 { 00:12:08.124 "name": null, 00:12:08.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.124 "is_configured": false, 00:12:08.124 "data_offset": 0, 00:12:08.124 "data_size": 65536 00:12:08.124 }, 00:12:08.124 { 00:12:08.124 "name": "BaseBdev2", 00:12:08.124 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:08.124 "is_configured": true, 00:12:08.124 "data_offset": 0, 00:12:08.124 "data_size": 65536 00:12:08.124 }, 00:12:08.124 { 00:12:08.124 "name": "BaseBdev3", 00:12:08.124 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:08.124 "is_configured": true, 00:12:08.124 "data_offset": 0, 00:12:08.124 "data_size": 65536 00:12:08.124 }, 00:12:08.124 { 00:12:08.124 "name": "BaseBdev4", 00:12:08.124 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:08.124 "is_configured": true, 00:12:08.124 "data_offset": 0, 00:12:08.124 "data_size": 65536 00:12:08.124 } 00:12:08.124 ] 00:12:08.124 }' 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.124 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.384 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:08.384 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.384 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.384 [2024-09-30 14:11:12.849801] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:08.384 [2024-09-30 14:11:12.853244] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09e40 00:12:08.384 [2024-09-30 14:11:12.855148] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:08.384 14:11:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.384 14:11:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:09.322 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:09.322 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.322 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:09.322 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.323 "name": "raid_bdev1", 00:12:09.323 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:09.323 "strip_size_kb": 0, 00:12:09.323 "state": "online", 00:12:09.323 "raid_level": "raid1", 00:12:09.323 "superblock": false, 00:12:09.323 "num_base_bdevs": 4, 00:12:09.323 "num_base_bdevs_discovered": 4, 00:12:09.323 "num_base_bdevs_operational": 4, 00:12:09.323 "process": { 00:12:09.323 "type": "rebuild", 00:12:09.323 "target": "spare", 00:12:09.323 "progress": { 00:12:09.323 "blocks": 20480, 00:12:09.323 "percent": 31 00:12:09.323 } 00:12:09.323 }, 00:12:09.323 "base_bdevs_list": [ 00:12:09.323 { 00:12:09.323 "name": "spare", 00:12:09.323 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:09.323 "is_configured": true, 00:12:09.323 "data_offset": 0, 00:12:09.323 "data_size": 65536 00:12:09.323 }, 00:12:09.323 { 00:12:09.323 "name": "BaseBdev2", 00:12:09.323 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:09.323 "is_configured": true, 00:12:09.323 "data_offset": 0, 00:12:09.323 "data_size": 65536 00:12:09.323 }, 00:12:09.323 { 00:12:09.323 "name": "BaseBdev3", 00:12:09.323 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:09.323 "is_configured": true, 00:12:09.323 "data_offset": 0, 00:12:09.323 "data_size": 65536 00:12:09.323 }, 00:12:09.323 { 00:12:09.323 "name": "BaseBdev4", 00:12:09.323 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:09.323 "is_configured": true, 00:12:09.323 "data_offset": 0, 00:12:09.323 "data_size": 65536 00:12:09.323 } 00:12:09.323 ] 00:12:09.323 }' 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:09.323 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:09.581 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:09.581 14:11:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:09.582 14:11:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.582 14:11:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.582 [2024-09-30 14:11:14.002126] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:09.582 [2024-09-30 14:11:14.060053] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:09.582 [2024-09-30 14:11:14.060116] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:09.582 [2024-09-30 14:11:14.060133] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:09.582 [2024-09-30 14:11:14.060144] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.582 "name": "raid_bdev1", 00:12:09.582 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:09.582 "strip_size_kb": 0, 00:12:09.582 "state": "online", 00:12:09.582 "raid_level": "raid1", 00:12:09.582 "superblock": false, 00:12:09.582 "num_base_bdevs": 4, 00:12:09.582 "num_base_bdevs_discovered": 3, 00:12:09.582 "num_base_bdevs_operational": 3, 00:12:09.582 "base_bdevs_list": [ 00:12:09.582 { 00:12:09.582 "name": null, 00:12:09.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.582 "is_configured": false, 00:12:09.582 "data_offset": 0, 00:12:09.582 "data_size": 65536 00:12:09.582 }, 00:12:09.582 { 00:12:09.582 "name": "BaseBdev2", 00:12:09.582 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:09.582 "is_configured": true, 00:12:09.582 "data_offset": 0, 00:12:09.582 "data_size": 65536 00:12:09.582 }, 00:12:09.582 { 00:12:09.582 "name": "BaseBdev3", 00:12:09.582 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:09.582 "is_configured": true, 00:12:09.582 "data_offset": 0, 00:12:09.582 "data_size": 65536 00:12:09.582 }, 00:12:09.582 { 00:12:09.582 "name": "BaseBdev4", 00:12:09.582 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:09.582 "is_configured": true, 00:12:09.582 "data_offset": 0, 00:12:09.582 "data_size": 65536 00:12:09.582 } 00:12:09.582 ] 00:12:09.582 }' 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.582 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:10.151 "name": "raid_bdev1", 00:12:10.151 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:10.151 "strip_size_kb": 0, 00:12:10.151 "state": "online", 00:12:10.151 "raid_level": "raid1", 00:12:10.151 "superblock": false, 00:12:10.151 "num_base_bdevs": 4, 00:12:10.151 "num_base_bdevs_discovered": 3, 00:12:10.151 "num_base_bdevs_operational": 3, 00:12:10.151 "base_bdevs_list": [ 00:12:10.151 { 00:12:10.151 "name": null, 00:12:10.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:10.151 "is_configured": false, 00:12:10.151 "data_offset": 0, 00:12:10.151 "data_size": 65536 00:12:10.151 }, 00:12:10.151 { 00:12:10.151 "name": "BaseBdev2", 00:12:10.151 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:10.151 "is_configured": true, 00:12:10.151 "data_offset": 0, 00:12:10.151 "data_size": 65536 00:12:10.151 }, 00:12:10.151 { 00:12:10.151 "name": "BaseBdev3", 00:12:10.151 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:10.151 "is_configured": true, 00:12:10.151 "data_offset": 0, 00:12:10.151 "data_size": 65536 00:12:10.151 }, 00:12:10.151 { 00:12:10.151 "name": "BaseBdev4", 00:12:10.151 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:10.151 "is_configured": true, 00:12:10.151 "data_offset": 0, 00:12:10.151 "data_size": 65536 00:12:10.151 } 00:12:10.151 ] 00:12:10.151 }' 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.151 [2024-09-30 14:11:14.626930] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:10.151 [2024-09-30 14:11:14.630053] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09f10 00:12:10.151 [2024-09-30 14:11:14.631933] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.151 14:11:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.088 "name": "raid_bdev1", 00:12:11.088 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:11.088 "strip_size_kb": 0, 00:12:11.088 "state": "online", 00:12:11.088 "raid_level": "raid1", 00:12:11.088 "superblock": false, 00:12:11.088 "num_base_bdevs": 4, 00:12:11.088 "num_base_bdevs_discovered": 4, 00:12:11.088 "num_base_bdevs_operational": 4, 00:12:11.088 "process": { 00:12:11.088 "type": "rebuild", 00:12:11.088 "target": "spare", 00:12:11.088 "progress": { 00:12:11.088 "blocks": 20480, 00:12:11.088 "percent": 31 00:12:11.088 } 00:12:11.088 }, 00:12:11.088 "base_bdevs_list": [ 00:12:11.088 { 00:12:11.088 "name": "spare", 00:12:11.088 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:11.088 "is_configured": true, 00:12:11.088 "data_offset": 0, 00:12:11.088 "data_size": 65536 00:12:11.088 }, 00:12:11.088 { 00:12:11.088 "name": "BaseBdev2", 00:12:11.088 "uuid": "67e29857-1487-5a07-8f29-d98026e247be", 00:12:11.088 "is_configured": true, 00:12:11.088 "data_offset": 0, 00:12:11.088 "data_size": 65536 00:12:11.088 }, 00:12:11.088 { 00:12:11.088 "name": "BaseBdev3", 00:12:11.088 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:11.088 "is_configured": true, 00:12:11.088 "data_offset": 0, 00:12:11.088 "data_size": 65536 00:12:11.088 }, 00:12:11.088 { 00:12:11.088 "name": "BaseBdev4", 00:12:11.088 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:11.088 "is_configured": true, 00:12:11.088 "data_offset": 0, 00:12:11.088 "data_size": 65536 00:12:11.088 } 00:12:11.088 ] 00:12:11.088 }' 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:11.088 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.348 [2024-09-30 14:11:15.790948] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:11.348 [2024-09-30 14:11:15.835733] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09f10 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.348 "name": "raid_bdev1", 00:12:11.348 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:11.348 "strip_size_kb": 0, 00:12:11.348 "state": "online", 00:12:11.348 "raid_level": "raid1", 00:12:11.348 "superblock": false, 00:12:11.348 "num_base_bdevs": 4, 00:12:11.348 "num_base_bdevs_discovered": 3, 00:12:11.348 "num_base_bdevs_operational": 3, 00:12:11.348 "process": { 00:12:11.348 "type": "rebuild", 00:12:11.348 "target": "spare", 00:12:11.348 "progress": { 00:12:11.348 "blocks": 24576, 00:12:11.348 "percent": 37 00:12:11.348 } 00:12:11.348 }, 00:12:11.348 "base_bdevs_list": [ 00:12:11.348 { 00:12:11.348 "name": "spare", 00:12:11.348 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:11.348 "is_configured": true, 00:12:11.348 "data_offset": 0, 00:12:11.348 "data_size": 65536 00:12:11.348 }, 00:12:11.348 { 00:12:11.348 "name": null, 00:12:11.348 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.348 "is_configured": false, 00:12:11.348 "data_offset": 0, 00:12:11.348 "data_size": 65536 00:12:11.348 }, 00:12:11.348 { 00:12:11.348 "name": "BaseBdev3", 00:12:11.348 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:11.348 "is_configured": true, 00:12:11.348 "data_offset": 0, 00:12:11.348 "data_size": 65536 00:12:11.348 }, 00:12:11.348 { 00:12:11.348 "name": "BaseBdev4", 00:12:11.348 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:11.348 "is_configured": true, 00:12:11.348 "data_offset": 0, 00:12:11.348 "data_size": 65536 00:12:11.348 } 00:12:11.348 ] 00:12:11.348 }' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=356 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.348 14:11:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.608 "name": "raid_bdev1", 00:12:11.608 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:11.608 "strip_size_kb": 0, 00:12:11.608 "state": "online", 00:12:11.608 "raid_level": "raid1", 00:12:11.608 "superblock": false, 00:12:11.608 "num_base_bdevs": 4, 00:12:11.608 "num_base_bdevs_discovered": 3, 00:12:11.608 "num_base_bdevs_operational": 3, 00:12:11.608 "process": { 00:12:11.608 "type": "rebuild", 00:12:11.608 "target": "spare", 00:12:11.608 "progress": { 00:12:11.608 "blocks": 26624, 00:12:11.608 "percent": 40 00:12:11.608 } 00:12:11.608 }, 00:12:11.608 "base_bdevs_list": [ 00:12:11.608 { 00:12:11.608 "name": "spare", 00:12:11.608 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:11.608 "is_configured": true, 00:12:11.608 "data_offset": 0, 00:12:11.608 "data_size": 65536 00:12:11.608 }, 00:12:11.608 { 00:12:11.608 "name": null, 00:12:11.608 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.608 "is_configured": false, 00:12:11.608 "data_offset": 0, 00:12:11.608 "data_size": 65536 00:12:11.608 }, 00:12:11.608 { 00:12:11.608 "name": "BaseBdev3", 00:12:11.608 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:11.608 "is_configured": true, 00:12:11.608 "data_offset": 0, 00:12:11.608 "data_size": 65536 00:12:11.608 }, 00:12:11.608 { 00:12:11.608 "name": "BaseBdev4", 00:12:11.608 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:11.608 "is_configured": true, 00:12:11.608 "data_offset": 0, 00:12:11.608 "data_size": 65536 00:12:11.608 } 00:12:11.608 ] 00:12:11.608 }' 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:11.608 14:11:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:12.546 "name": "raid_bdev1", 00:12:12.546 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:12.546 "strip_size_kb": 0, 00:12:12.546 "state": "online", 00:12:12.546 "raid_level": "raid1", 00:12:12.546 "superblock": false, 00:12:12.546 "num_base_bdevs": 4, 00:12:12.546 "num_base_bdevs_discovered": 3, 00:12:12.546 "num_base_bdevs_operational": 3, 00:12:12.546 "process": { 00:12:12.546 "type": "rebuild", 00:12:12.546 "target": "spare", 00:12:12.546 "progress": { 00:12:12.546 "blocks": 49152, 00:12:12.546 "percent": 75 00:12:12.546 } 00:12:12.546 }, 00:12:12.546 "base_bdevs_list": [ 00:12:12.546 { 00:12:12.546 "name": "spare", 00:12:12.546 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 0, 00:12:12.546 "data_size": 65536 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": null, 00:12:12.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:12.546 "is_configured": false, 00:12:12.546 "data_offset": 0, 00:12:12.546 "data_size": 65536 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": "BaseBdev3", 00:12:12.546 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 0, 00:12:12.546 "data_size": 65536 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": "BaseBdev4", 00:12:12.546 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 0, 00:12:12.546 "data_size": 65536 00:12:12.546 } 00:12:12.546 ] 00:12:12.546 }' 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:12.546 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:12.805 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:12.805 14:11:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:13.374 [2024-09-30 14:11:17.842370] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:13.374 [2024-09-30 14:11:17.842432] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:13.374 [2024-09-30 14:11:17.842477] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.633 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.893 "name": "raid_bdev1", 00:12:13.893 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:13.893 "strip_size_kb": 0, 00:12:13.893 "state": "online", 00:12:13.893 "raid_level": "raid1", 00:12:13.893 "superblock": false, 00:12:13.893 "num_base_bdevs": 4, 00:12:13.893 "num_base_bdevs_discovered": 3, 00:12:13.893 "num_base_bdevs_operational": 3, 00:12:13.893 "base_bdevs_list": [ 00:12:13.893 { 00:12:13.893 "name": "spare", 00:12:13.893 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": null, 00:12:13.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.893 "is_configured": false, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": "BaseBdev3", 00:12:13.893 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": "BaseBdev4", 00:12:13.893 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 } 00:12:13.893 ] 00:12:13.893 }' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.893 "name": "raid_bdev1", 00:12:13.893 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:13.893 "strip_size_kb": 0, 00:12:13.893 "state": "online", 00:12:13.893 "raid_level": "raid1", 00:12:13.893 "superblock": false, 00:12:13.893 "num_base_bdevs": 4, 00:12:13.893 "num_base_bdevs_discovered": 3, 00:12:13.893 "num_base_bdevs_operational": 3, 00:12:13.893 "base_bdevs_list": [ 00:12:13.893 { 00:12:13.893 "name": "spare", 00:12:13.893 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": null, 00:12:13.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.893 "is_configured": false, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": "BaseBdev3", 00:12:13.893 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 }, 00:12:13.893 { 00:12:13.893 "name": "BaseBdev4", 00:12:13.893 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:13.893 "is_configured": true, 00:12:13.893 "data_offset": 0, 00:12:13.893 "data_size": 65536 00:12:13.893 } 00:12:13.893 ] 00:12:13.893 }' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:13.893 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.156 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:14.156 "name": "raid_bdev1", 00:12:14.156 "uuid": "f77736ff-8b97-43b3-ba48-a82f098db4ec", 00:12:14.156 "strip_size_kb": 0, 00:12:14.156 "state": "online", 00:12:14.157 "raid_level": "raid1", 00:12:14.157 "superblock": false, 00:12:14.157 "num_base_bdevs": 4, 00:12:14.157 "num_base_bdevs_discovered": 3, 00:12:14.157 "num_base_bdevs_operational": 3, 00:12:14.157 "base_bdevs_list": [ 00:12:14.157 { 00:12:14.157 "name": "spare", 00:12:14.157 "uuid": "9649216a-05bb-5b9c-b21b-7dc4a0e5f0ef", 00:12:14.157 "is_configured": true, 00:12:14.157 "data_offset": 0, 00:12:14.157 "data_size": 65536 00:12:14.157 }, 00:12:14.157 { 00:12:14.157 "name": null, 00:12:14.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:14.157 "is_configured": false, 00:12:14.157 "data_offset": 0, 00:12:14.157 "data_size": 65536 00:12:14.157 }, 00:12:14.157 { 00:12:14.157 "name": "BaseBdev3", 00:12:14.157 "uuid": "733e0dff-b6a8-5797-a473-bf533521c7e5", 00:12:14.157 "is_configured": true, 00:12:14.157 "data_offset": 0, 00:12:14.157 "data_size": 65536 00:12:14.157 }, 00:12:14.157 { 00:12:14.157 "name": "BaseBdev4", 00:12:14.157 "uuid": "6f5fe256-7fb4-5cc0-889b-e4f39b305cf7", 00:12:14.157 "is_configured": true, 00:12:14.157 "data_offset": 0, 00:12:14.157 "data_size": 65536 00:12:14.157 } 00:12:14.157 ] 00:12:14.157 }' 00:12:14.157 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:14.157 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.423 14:11:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:14.423 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.423 14:11:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.423 [2024-09-30 14:11:19.003995] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:14.423 [2024-09-30 14:11:19.004026] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:14.423 [2024-09-30 14:11:19.004099] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:14.423 [2024-09-30 14:11:19.004178] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:14.423 [2024-09-30 14:11:19.004187] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:14.423 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:14.683 /dev/nbd0 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:14.683 1+0 records in 00:12:14.683 1+0 records out 00:12:14.683 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000378022 s, 10.8 MB/s 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:14.683 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:14.942 /dev/nbd1 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:14.942 1+0 records in 00:12:14.942 1+0 records out 00:12:14.942 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000394642 s, 10.4 MB/s 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:14.942 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:15.202 14:11:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:15.202 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:15.203 14:11:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 89220 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 89220 ']' 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 89220 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89220 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:15.463 killing process with pid 89220 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89220' 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 89220 00:12:15.463 Received shutdown signal, test time was about 60.000000 seconds 00:12:15.463 00:12:15.463 Latency(us) 00:12:15.463 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:15.463 =================================================================================================================== 00:12:15.463 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:15.463 [2024-09-30 14:11:20.093301] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:15.463 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 89220 00:12:15.723 [2024-09-30 14:11:20.142363] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:15.983 00:12:15.983 real 0m14.987s 00:12:15.983 user 0m17.199s 00:12:15.983 sys 0m2.824s 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:15.983 ************************************ 00:12:15.983 END TEST raid_rebuild_test 00:12:15.983 ************************************ 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.983 14:11:20 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:15.983 14:11:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:15.983 14:11:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:15.983 14:11:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:15.983 ************************************ 00:12:15.983 START TEST raid_rebuild_test_sb 00:12:15.983 ************************************ 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true false true 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=89640 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 89640 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 89640 ']' 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:15.983 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:15.983 14:11:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:15.983 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:15.983 Zero copy mechanism will not be used. 00:12:15.983 [2024-09-30 14:11:20.555883] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:12:15.984 [2024-09-30 14:11:20.556067] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89640 ] 00:12:16.244 [2024-09-30 14:11:20.686602] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:12:16.244 [2024-09-30 14:11:20.716628] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:16.244 [2024-09-30 14:11:20.761220] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:16.244 [2024-09-30 14:11:20.801819] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:16.244 [2024-09-30 14:11:20.801863] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.815 BaseBdev1_malloc 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.815 [2024-09-30 14:11:21.414754] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:16.815 [2024-09-30 14:11:21.414915] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.815 [2024-09-30 14:11:21.414960] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:16.815 [2024-09-30 14:11:21.415021] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.815 [2024-09-30 14:11:21.417079] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.815 [2024-09-30 14:11:21.417148] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:16.815 BaseBdev1 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.815 BaseBdev2_malloc 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.815 [2024-09-30 14:11:21.460340] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:16.815 [2024-09-30 14:11:21.460678] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.815 [2024-09-30 14:11:21.460827] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:16.815 [2024-09-30 14:11:21.460935] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.815 [2024-09-30 14:11:21.465429] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.815 [2024-09-30 14:11:21.465601] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:16.815 BaseBdev2 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.815 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 BaseBdev3_malloc 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 [2024-09-30 14:11:21.491441] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:17.076 [2024-09-30 14:11:21.491580] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:17.076 [2024-09-30 14:11:21.491636] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:17.076 [2024-09-30 14:11:21.491671] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:17.076 [2024-09-30 14:11:21.493685] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:17.076 [2024-09-30 14:11:21.493755] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:17.076 BaseBdev3 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 BaseBdev4_malloc 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 [2024-09-30 14:11:21.519665] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:17.076 [2024-09-30 14:11:21.519771] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:17.076 [2024-09-30 14:11:21.519821] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:17.076 [2024-09-30 14:11:21.519855] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:17.076 [2024-09-30 14:11:21.521880] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:17.076 [2024-09-30 14:11:21.521949] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:17.076 BaseBdev4 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 spare_malloc 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 spare_delay 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 [2024-09-30 14:11:21.560062] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:17.076 [2024-09-30 14:11:21.560173] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:17.076 [2024-09-30 14:11:21.560211] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:17.076 [2024-09-30 14:11:21.560241] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:17.076 [2024-09-30 14:11:21.562261] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:17.076 [2024-09-30 14:11:21.562334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:17.076 spare 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.076 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.076 [2024-09-30 14:11:21.572119] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:17.076 [2024-09-30 14:11:21.573831] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:17.076 [2024-09-30 14:11:21.573891] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:17.076 [2024-09-30 14:11:21.573934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:17.076 [2024-09-30 14:11:21.574080] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:17.076 [2024-09-30 14:11:21.574097] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:17.077 [2024-09-30 14:11:21.574312] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:17.077 [2024-09-30 14:11:21.574431] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:17.077 [2024-09-30 14:11:21.574440] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:17.077 [2024-09-30 14:11:21.574568] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:17.077 "name": "raid_bdev1", 00:12:17.077 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:17.077 "strip_size_kb": 0, 00:12:17.077 "state": "online", 00:12:17.077 "raid_level": "raid1", 00:12:17.077 "superblock": true, 00:12:17.077 "num_base_bdevs": 4, 00:12:17.077 "num_base_bdevs_discovered": 4, 00:12:17.077 "num_base_bdevs_operational": 4, 00:12:17.077 "base_bdevs_list": [ 00:12:17.077 { 00:12:17.077 "name": "BaseBdev1", 00:12:17.077 "uuid": "7f17f89e-e412-5b1e-b187-dca3f50d9c50", 00:12:17.077 "is_configured": true, 00:12:17.077 "data_offset": 2048, 00:12:17.077 "data_size": 63488 00:12:17.077 }, 00:12:17.077 { 00:12:17.077 "name": "BaseBdev2", 00:12:17.077 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:17.077 "is_configured": true, 00:12:17.077 "data_offset": 2048, 00:12:17.077 "data_size": 63488 00:12:17.077 }, 00:12:17.077 { 00:12:17.077 "name": "BaseBdev3", 00:12:17.077 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:17.077 "is_configured": true, 00:12:17.077 "data_offset": 2048, 00:12:17.077 "data_size": 63488 00:12:17.077 }, 00:12:17.077 { 00:12:17.077 "name": "BaseBdev4", 00:12:17.077 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:17.077 "is_configured": true, 00:12:17.077 "data_offset": 2048, 00:12:17.077 "data_size": 63488 00:12:17.077 } 00:12:17.077 ] 00:12:17.077 }' 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:17.077 14:11:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.646 [2024-09-30 14:11:22.023622] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:17.646 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:17.646 [2024-09-30 14:11:22.283021] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:12:17.646 /dev/nbd0 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:17.959 1+0 records in 00:12:17.959 1+0 records out 00:12:17.959 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000570399 s, 7.2 MB/s 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:17.959 14:11:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:17.960 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:17.960 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:17.960 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:17.960 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:17.960 14:11:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:23.235 63488+0 records in 00:12:23.236 63488+0 records out 00:12:23.236 32505856 bytes (33 MB, 31 MiB) copied, 5.00191 s, 6.5 MB/s 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:23.236 [2024-09-30 14:11:27.546829] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.236 [2024-09-30 14:11:27.579339] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:23.236 "name": "raid_bdev1", 00:12:23.236 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:23.236 "strip_size_kb": 0, 00:12:23.236 "state": "online", 00:12:23.236 "raid_level": "raid1", 00:12:23.236 "superblock": true, 00:12:23.236 "num_base_bdevs": 4, 00:12:23.236 "num_base_bdevs_discovered": 3, 00:12:23.236 "num_base_bdevs_operational": 3, 00:12:23.236 "base_bdevs_list": [ 00:12:23.236 { 00:12:23.236 "name": null, 00:12:23.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.236 "is_configured": false, 00:12:23.236 "data_offset": 0, 00:12:23.236 "data_size": 63488 00:12:23.236 }, 00:12:23.236 { 00:12:23.236 "name": "BaseBdev2", 00:12:23.236 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:23.236 "is_configured": true, 00:12:23.236 "data_offset": 2048, 00:12:23.236 "data_size": 63488 00:12:23.236 }, 00:12:23.236 { 00:12:23.236 "name": "BaseBdev3", 00:12:23.236 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:23.236 "is_configured": true, 00:12:23.236 "data_offset": 2048, 00:12:23.236 "data_size": 63488 00:12:23.236 }, 00:12:23.236 { 00:12:23.236 "name": "BaseBdev4", 00:12:23.236 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:23.236 "is_configured": true, 00:12:23.236 "data_offset": 2048, 00:12:23.236 "data_size": 63488 00:12:23.236 } 00:12:23.236 ] 00:12:23.236 }' 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:23.236 14:11:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.495 14:11:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:23.495 14:11:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.495 14:11:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.495 [2024-09-30 14:11:28.030718] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:23.495 [2024-09-30 14:11:28.034119] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca35d0 00:12:23.495 [2024-09-30 14:11:28.035952] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:23.495 14:11:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.495 14:11:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.433 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.434 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.434 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.434 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.434 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.693 "name": "raid_bdev1", 00:12:24.693 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:24.693 "strip_size_kb": 0, 00:12:24.693 "state": "online", 00:12:24.693 "raid_level": "raid1", 00:12:24.693 "superblock": true, 00:12:24.693 "num_base_bdevs": 4, 00:12:24.693 "num_base_bdevs_discovered": 4, 00:12:24.693 "num_base_bdevs_operational": 4, 00:12:24.693 "process": { 00:12:24.693 "type": "rebuild", 00:12:24.693 "target": "spare", 00:12:24.693 "progress": { 00:12:24.693 "blocks": 20480, 00:12:24.693 "percent": 32 00:12:24.693 } 00:12:24.693 }, 00:12:24.693 "base_bdevs_list": [ 00:12:24.693 { 00:12:24.693 "name": "spare", 00:12:24.693 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:24.693 "is_configured": true, 00:12:24.693 "data_offset": 2048, 00:12:24.693 "data_size": 63488 00:12:24.693 }, 00:12:24.693 { 00:12:24.693 "name": "BaseBdev2", 00:12:24.693 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:24.693 "is_configured": true, 00:12:24.693 "data_offset": 2048, 00:12:24.693 "data_size": 63488 00:12:24.693 }, 00:12:24.693 { 00:12:24.693 "name": "BaseBdev3", 00:12:24.693 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:24.693 "is_configured": true, 00:12:24.693 "data_offset": 2048, 00:12:24.693 "data_size": 63488 00:12:24.693 }, 00:12:24.693 { 00:12:24.693 "name": "BaseBdev4", 00:12:24.693 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:24.693 "is_configured": true, 00:12:24.693 "data_offset": 2048, 00:12:24.693 "data_size": 63488 00:12:24.693 } 00:12:24.693 ] 00:12:24.693 }' 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.693 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.693 [2024-09-30 14:11:29.179728] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:24.693 [2024-09-30 14:11:29.240704] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:24.694 [2024-09-30 14:11:29.240782] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:24.694 [2024-09-30 14:11:29.240815] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:24.694 [2024-09-30 14:11:29.240827] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:24.694 "name": "raid_bdev1", 00:12:24.694 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:24.694 "strip_size_kb": 0, 00:12:24.694 "state": "online", 00:12:24.694 "raid_level": "raid1", 00:12:24.694 "superblock": true, 00:12:24.694 "num_base_bdevs": 4, 00:12:24.694 "num_base_bdevs_discovered": 3, 00:12:24.694 "num_base_bdevs_operational": 3, 00:12:24.694 "base_bdevs_list": [ 00:12:24.694 { 00:12:24.694 "name": null, 00:12:24.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.694 "is_configured": false, 00:12:24.694 "data_offset": 0, 00:12:24.694 "data_size": 63488 00:12:24.694 }, 00:12:24.694 { 00:12:24.694 "name": "BaseBdev2", 00:12:24.694 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:24.694 "is_configured": true, 00:12:24.694 "data_offset": 2048, 00:12:24.694 "data_size": 63488 00:12:24.694 }, 00:12:24.694 { 00:12:24.694 "name": "BaseBdev3", 00:12:24.694 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:24.694 "is_configured": true, 00:12:24.694 "data_offset": 2048, 00:12:24.694 "data_size": 63488 00:12:24.694 }, 00:12:24.694 { 00:12:24.694 "name": "BaseBdev4", 00:12:24.694 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:24.694 "is_configured": true, 00:12:24.694 "data_offset": 2048, 00:12:24.694 "data_size": 63488 00:12:24.694 } 00:12:24.694 ] 00:12:24.694 }' 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:24.694 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.952 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.211 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:25.212 "name": "raid_bdev1", 00:12:25.212 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:25.212 "strip_size_kb": 0, 00:12:25.212 "state": "online", 00:12:25.212 "raid_level": "raid1", 00:12:25.212 "superblock": true, 00:12:25.212 "num_base_bdevs": 4, 00:12:25.212 "num_base_bdevs_discovered": 3, 00:12:25.212 "num_base_bdevs_operational": 3, 00:12:25.212 "base_bdevs_list": [ 00:12:25.212 { 00:12:25.212 "name": null, 00:12:25.212 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.212 "is_configured": false, 00:12:25.212 "data_offset": 0, 00:12:25.212 "data_size": 63488 00:12:25.212 }, 00:12:25.212 { 00:12:25.212 "name": "BaseBdev2", 00:12:25.212 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:25.212 "is_configured": true, 00:12:25.212 "data_offset": 2048, 00:12:25.212 "data_size": 63488 00:12:25.212 }, 00:12:25.212 { 00:12:25.212 "name": "BaseBdev3", 00:12:25.212 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:25.212 "is_configured": true, 00:12:25.212 "data_offset": 2048, 00:12:25.212 "data_size": 63488 00:12:25.212 }, 00:12:25.212 { 00:12:25.212 "name": "BaseBdev4", 00:12:25.212 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:25.212 "is_configured": true, 00:12:25.212 "data_offset": 2048, 00:12:25.212 "data_size": 63488 00:12:25.212 } 00:12:25.212 ] 00:12:25.212 }' 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.212 [2024-09-30 14:11:29.712221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:25.212 [2024-09-30 14:11:29.715576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca36a0 00:12:25.212 [2024-09-30 14:11:29.717413] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.212 14:11:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:26.149 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.149 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.150 "name": "raid_bdev1", 00:12:26.150 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:26.150 "strip_size_kb": 0, 00:12:26.150 "state": "online", 00:12:26.150 "raid_level": "raid1", 00:12:26.150 "superblock": true, 00:12:26.150 "num_base_bdevs": 4, 00:12:26.150 "num_base_bdevs_discovered": 4, 00:12:26.150 "num_base_bdevs_operational": 4, 00:12:26.150 "process": { 00:12:26.150 "type": "rebuild", 00:12:26.150 "target": "spare", 00:12:26.150 "progress": { 00:12:26.150 "blocks": 20480, 00:12:26.150 "percent": 32 00:12:26.150 } 00:12:26.150 }, 00:12:26.150 "base_bdevs_list": [ 00:12:26.150 { 00:12:26.150 "name": "spare", 00:12:26.150 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:26.150 "is_configured": true, 00:12:26.150 "data_offset": 2048, 00:12:26.150 "data_size": 63488 00:12:26.150 }, 00:12:26.150 { 00:12:26.150 "name": "BaseBdev2", 00:12:26.150 "uuid": "e84bc6ff-c7c1-500a-b9ef-bc5cc727c5dc", 00:12:26.150 "is_configured": true, 00:12:26.150 "data_offset": 2048, 00:12:26.150 "data_size": 63488 00:12:26.150 }, 00:12:26.150 { 00:12:26.150 "name": "BaseBdev3", 00:12:26.150 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:26.150 "is_configured": true, 00:12:26.150 "data_offset": 2048, 00:12:26.150 "data_size": 63488 00:12:26.150 }, 00:12:26.150 { 00:12:26.150 "name": "BaseBdev4", 00:12:26.150 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:26.150 "is_configured": true, 00:12:26.150 "data_offset": 2048, 00:12:26.150 "data_size": 63488 00:12:26.150 } 00:12:26.150 ] 00:12:26.150 }' 00:12:26.150 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:26.409 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.409 14:11:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.409 [2024-09-30 14:11:30.864665] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:26.409 [2024-09-30 14:11:31.021333] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca36a0 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.409 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.410 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.410 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.410 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.669 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.669 "name": "raid_bdev1", 00:12:26.669 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:26.669 "strip_size_kb": 0, 00:12:26.669 "state": "online", 00:12:26.669 "raid_level": "raid1", 00:12:26.669 "superblock": true, 00:12:26.669 "num_base_bdevs": 4, 00:12:26.669 "num_base_bdevs_discovered": 3, 00:12:26.669 "num_base_bdevs_operational": 3, 00:12:26.669 "process": { 00:12:26.669 "type": "rebuild", 00:12:26.669 "target": "spare", 00:12:26.669 "progress": { 00:12:26.669 "blocks": 24576, 00:12:26.669 "percent": 38 00:12:26.669 } 00:12:26.669 }, 00:12:26.669 "base_bdevs_list": [ 00:12:26.669 { 00:12:26.669 "name": "spare", 00:12:26.669 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": null, 00:12:26.670 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.670 "is_configured": false, 00:12:26.670 "data_offset": 0, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": "BaseBdev3", 00:12:26.670 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": "BaseBdev4", 00:12:26.670 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 } 00:12:26.670 ] 00:12:26.670 }' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=372 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.670 "name": "raid_bdev1", 00:12:26.670 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:26.670 "strip_size_kb": 0, 00:12:26.670 "state": "online", 00:12:26.670 "raid_level": "raid1", 00:12:26.670 "superblock": true, 00:12:26.670 "num_base_bdevs": 4, 00:12:26.670 "num_base_bdevs_discovered": 3, 00:12:26.670 "num_base_bdevs_operational": 3, 00:12:26.670 "process": { 00:12:26.670 "type": "rebuild", 00:12:26.670 "target": "spare", 00:12:26.670 "progress": { 00:12:26.670 "blocks": 26624, 00:12:26.670 "percent": 41 00:12:26.670 } 00:12:26.670 }, 00:12:26.670 "base_bdevs_list": [ 00:12:26.670 { 00:12:26.670 "name": "spare", 00:12:26.670 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": null, 00:12:26.670 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.670 "is_configured": false, 00:12:26.670 "data_offset": 0, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": "BaseBdev3", 00:12:26.670 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 }, 00:12:26.670 { 00:12:26.670 "name": "BaseBdev4", 00:12:26.670 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:26.670 "is_configured": true, 00:12:26.670 "data_offset": 2048, 00:12:26.670 "data_size": 63488 00:12:26.670 } 00:12:26.670 ] 00:12:26.670 }' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.670 14:11:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:28.052 "name": "raid_bdev1", 00:12:28.052 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:28.052 "strip_size_kb": 0, 00:12:28.052 "state": "online", 00:12:28.052 "raid_level": "raid1", 00:12:28.052 "superblock": true, 00:12:28.052 "num_base_bdevs": 4, 00:12:28.052 "num_base_bdevs_discovered": 3, 00:12:28.052 "num_base_bdevs_operational": 3, 00:12:28.052 "process": { 00:12:28.052 "type": "rebuild", 00:12:28.052 "target": "spare", 00:12:28.052 "progress": { 00:12:28.052 "blocks": 51200, 00:12:28.052 "percent": 80 00:12:28.052 } 00:12:28.052 }, 00:12:28.052 "base_bdevs_list": [ 00:12:28.052 { 00:12:28.052 "name": "spare", 00:12:28.052 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:28.052 "is_configured": true, 00:12:28.052 "data_offset": 2048, 00:12:28.052 "data_size": 63488 00:12:28.052 }, 00:12:28.052 { 00:12:28.052 "name": null, 00:12:28.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:28.052 "is_configured": false, 00:12:28.052 "data_offset": 0, 00:12:28.052 "data_size": 63488 00:12:28.052 }, 00:12:28.052 { 00:12:28.052 "name": "BaseBdev3", 00:12:28.052 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:28.052 "is_configured": true, 00:12:28.052 "data_offset": 2048, 00:12:28.052 "data_size": 63488 00:12:28.052 }, 00:12:28.052 { 00:12:28.052 "name": "BaseBdev4", 00:12:28.052 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:28.052 "is_configured": true, 00:12:28.052 "data_offset": 2048, 00:12:28.052 "data_size": 63488 00:12:28.052 } 00:12:28.052 ] 00:12:28.052 }' 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:28.052 14:11:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:28.312 [2024-09-30 14:11:32.928179] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:28.312 [2024-09-30 14:11:32.928241] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:28.312 [2024-09-30 14:11:32.928340] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:28.881 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:28.882 "name": "raid_bdev1", 00:12:28.882 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:28.882 "strip_size_kb": 0, 00:12:28.882 "state": "online", 00:12:28.882 "raid_level": "raid1", 00:12:28.882 "superblock": true, 00:12:28.882 "num_base_bdevs": 4, 00:12:28.882 "num_base_bdevs_discovered": 3, 00:12:28.882 "num_base_bdevs_operational": 3, 00:12:28.882 "base_bdevs_list": [ 00:12:28.882 { 00:12:28.882 "name": "spare", 00:12:28.882 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:28.882 "is_configured": true, 00:12:28.882 "data_offset": 2048, 00:12:28.882 "data_size": 63488 00:12:28.882 }, 00:12:28.882 { 00:12:28.882 "name": null, 00:12:28.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:28.882 "is_configured": false, 00:12:28.882 "data_offset": 0, 00:12:28.882 "data_size": 63488 00:12:28.882 }, 00:12:28.882 { 00:12:28.882 "name": "BaseBdev3", 00:12:28.882 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:28.882 "is_configured": true, 00:12:28.882 "data_offset": 2048, 00:12:28.882 "data_size": 63488 00:12:28.882 }, 00:12:28.882 { 00:12:28.882 "name": "BaseBdev4", 00:12:28.882 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:28.882 "is_configured": true, 00:12:28.882 "data_offset": 2048, 00:12:28.882 "data_size": 63488 00:12:28.882 } 00:12:28.882 ] 00:12:28.882 }' 00:12:28.882 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.142 "name": "raid_bdev1", 00:12:29.142 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:29.142 "strip_size_kb": 0, 00:12:29.142 "state": "online", 00:12:29.142 "raid_level": "raid1", 00:12:29.142 "superblock": true, 00:12:29.142 "num_base_bdevs": 4, 00:12:29.142 "num_base_bdevs_discovered": 3, 00:12:29.142 "num_base_bdevs_operational": 3, 00:12:29.142 "base_bdevs_list": [ 00:12:29.142 { 00:12:29.142 "name": "spare", 00:12:29.142 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": null, 00:12:29.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.142 "is_configured": false, 00:12:29.142 "data_offset": 0, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": "BaseBdev3", 00:12:29.142 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": "BaseBdev4", 00:12:29.142 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 } 00:12:29.142 ] 00:12:29.142 }' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.142 "name": "raid_bdev1", 00:12:29.142 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:29.142 "strip_size_kb": 0, 00:12:29.142 "state": "online", 00:12:29.142 "raid_level": "raid1", 00:12:29.142 "superblock": true, 00:12:29.142 "num_base_bdevs": 4, 00:12:29.142 "num_base_bdevs_discovered": 3, 00:12:29.142 "num_base_bdevs_operational": 3, 00:12:29.142 "base_bdevs_list": [ 00:12:29.142 { 00:12:29.142 "name": "spare", 00:12:29.142 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": null, 00:12:29.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.142 "is_configured": false, 00:12:29.142 "data_offset": 0, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": "BaseBdev3", 00:12:29.142 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 }, 00:12:29.142 { 00:12:29.142 "name": "BaseBdev4", 00:12:29.142 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:29.142 "is_configured": true, 00:12:29.142 "data_offset": 2048, 00:12:29.142 "data_size": 63488 00:12:29.142 } 00:12:29.142 ] 00:12:29.142 }' 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.142 14:11:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.711 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:29.711 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.712 [2024-09-30 14:11:34.162355] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:29.712 [2024-09-30 14:11:34.162436] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:29.712 [2024-09-30 14:11:34.162567] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:29.712 [2024-09-30 14:11:34.162680] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:29.712 [2024-09-30 14:11:34.162731] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:29.712 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:29.972 /dev/nbd0 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:29.972 1+0 records in 00:12:29.972 1+0 records out 00:12:29.972 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000611004 s, 6.7 MB/s 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:29.972 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:30.232 /dev/nbd1 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:30.232 1+0 records in 00:12:30.232 1+0 records out 00:12:30.232 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000452086 s, 9.1 MB/s 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:30.232 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:30.492 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:30.492 14:11:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:30.492 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:30.752 [2024-09-30 14:11:35.238879] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:30.752 [2024-09-30 14:11:35.239001] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:30.752 [2024-09-30 14:11:35.239092] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:30.752 [2024-09-30 14:11:35.239132] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:30.752 [2024-09-30 14:11:35.241263] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:30.752 [2024-09-30 14:11:35.241334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:30.752 [2024-09-30 14:11:35.241436] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:30.752 [2024-09-30 14:11:35.241496] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.752 [2024-09-30 14:11:35.241628] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:30.752 [2024-09-30 14:11:35.241753] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:30.752 spare 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:30.752 [2024-09-30 14:11:35.341689] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:12:30.752 [2024-09-30 14:11:35.341757] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:30.752 [2024-09-30 14:11:35.342054] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1d50 00:12:30.752 [2024-09-30 14:11:35.342212] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:12:30.752 [2024-09-30 14:11:35.342255] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:12:30.752 [2024-09-30 14:11:35.342420] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:30.752 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:30.753 "name": "raid_bdev1", 00:12:30.753 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:30.753 "strip_size_kb": 0, 00:12:30.753 "state": "online", 00:12:30.753 "raid_level": "raid1", 00:12:30.753 "superblock": true, 00:12:30.753 "num_base_bdevs": 4, 00:12:30.753 "num_base_bdevs_discovered": 3, 00:12:30.753 "num_base_bdevs_operational": 3, 00:12:30.753 "base_bdevs_list": [ 00:12:30.753 { 00:12:30.753 "name": "spare", 00:12:30.753 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:30.753 "is_configured": true, 00:12:30.753 "data_offset": 2048, 00:12:30.753 "data_size": 63488 00:12:30.753 }, 00:12:30.753 { 00:12:30.753 "name": null, 00:12:30.753 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.753 "is_configured": false, 00:12:30.753 "data_offset": 2048, 00:12:30.753 "data_size": 63488 00:12:30.753 }, 00:12:30.753 { 00:12:30.753 "name": "BaseBdev3", 00:12:30.753 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:30.753 "is_configured": true, 00:12:30.753 "data_offset": 2048, 00:12:30.753 "data_size": 63488 00:12:30.753 }, 00:12:30.753 { 00:12:30.753 "name": "BaseBdev4", 00:12:30.753 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:30.753 "is_configured": true, 00:12:30.753 "data_offset": 2048, 00:12:30.753 "data_size": 63488 00:12:30.753 } 00:12:30.753 ] 00:12:30.753 }' 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:30.753 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.321 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.321 "name": "raid_bdev1", 00:12:31.321 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:31.321 "strip_size_kb": 0, 00:12:31.321 "state": "online", 00:12:31.321 "raid_level": "raid1", 00:12:31.322 "superblock": true, 00:12:31.322 "num_base_bdevs": 4, 00:12:31.322 "num_base_bdevs_discovered": 3, 00:12:31.322 "num_base_bdevs_operational": 3, 00:12:31.322 "base_bdevs_list": [ 00:12:31.322 { 00:12:31.322 "name": "spare", 00:12:31.322 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:31.322 "is_configured": true, 00:12:31.322 "data_offset": 2048, 00:12:31.322 "data_size": 63488 00:12:31.322 }, 00:12:31.322 { 00:12:31.322 "name": null, 00:12:31.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.322 "is_configured": false, 00:12:31.322 "data_offset": 2048, 00:12:31.322 "data_size": 63488 00:12:31.322 }, 00:12:31.322 { 00:12:31.322 "name": "BaseBdev3", 00:12:31.322 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:31.322 "is_configured": true, 00:12:31.322 "data_offset": 2048, 00:12:31.322 "data_size": 63488 00:12:31.322 }, 00:12:31.322 { 00:12:31.322 "name": "BaseBdev4", 00:12:31.322 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:31.322 "is_configured": true, 00:12:31.322 "data_offset": 2048, 00:12:31.322 "data_size": 63488 00:12:31.322 } 00:12:31.322 ] 00:12:31.322 }' 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.322 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.581 [2024-09-30 14:11:35.989616] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.581 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.582 14:11:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.582 "name": "raid_bdev1", 00:12:31.582 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:31.582 "strip_size_kb": 0, 00:12:31.582 "state": "online", 00:12:31.582 "raid_level": "raid1", 00:12:31.582 "superblock": true, 00:12:31.582 "num_base_bdevs": 4, 00:12:31.582 "num_base_bdevs_discovered": 2, 00:12:31.582 "num_base_bdevs_operational": 2, 00:12:31.582 "base_bdevs_list": [ 00:12:31.582 { 00:12:31.582 "name": null, 00:12:31.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.582 "is_configured": false, 00:12:31.582 "data_offset": 0, 00:12:31.582 "data_size": 63488 00:12:31.582 }, 00:12:31.582 { 00:12:31.582 "name": null, 00:12:31.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.582 "is_configured": false, 00:12:31.582 "data_offset": 2048, 00:12:31.582 "data_size": 63488 00:12:31.582 }, 00:12:31.582 { 00:12:31.582 "name": "BaseBdev3", 00:12:31.582 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:31.582 "is_configured": true, 00:12:31.582 "data_offset": 2048, 00:12:31.582 "data_size": 63488 00:12:31.582 }, 00:12:31.582 { 00:12:31.582 "name": "BaseBdev4", 00:12:31.582 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:31.582 "is_configured": true, 00:12:31.582 "data_offset": 2048, 00:12:31.582 "data_size": 63488 00:12:31.582 } 00:12:31.582 ] 00:12:31.582 }' 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.582 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.842 14:11:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:31.842 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.842 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.842 [2024-09-30 14:11:36.432886] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:31.842 [2024-09-30 14:11:36.433143] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:31.842 [2024-09-30 14:11:36.433208] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:31.842 [2024-09-30 14:11:36.433273] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:31.842 [2024-09-30 14:11:36.436405] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1e20 00:12:31.842 [2024-09-30 14:11:36.438282] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:31.842 14:11:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.842 14:11:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.224 "name": "raid_bdev1", 00:12:33.224 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:33.224 "strip_size_kb": 0, 00:12:33.224 "state": "online", 00:12:33.224 "raid_level": "raid1", 00:12:33.224 "superblock": true, 00:12:33.224 "num_base_bdevs": 4, 00:12:33.224 "num_base_bdevs_discovered": 3, 00:12:33.224 "num_base_bdevs_operational": 3, 00:12:33.224 "process": { 00:12:33.224 "type": "rebuild", 00:12:33.224 "target": "spare", 00:12:33.224 "progress": { 00:12:33.224 "blocks": 20480, 00:12:33.224 "percent": 32 00:12:33.224 } 00:12:33.224 }, 00:12:33.224 "base_bdevs_list": [ 00:12:33.224 { 00:12:33.224 "name": "spare", 00:12:33.224 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:33.224 "is_configured": true, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": null, 00:12:33.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.224 "is_configured": false, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": "BaseBdev3", 00:12:33.224 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:33.224 "is_configured": true, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": "BaseBdev4", 00:12:33.224 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:33.224 "is_configured": true, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 } 00:12:33.224 ] 00:12:33.224 }' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.224 [2024-09-30 14:11:37.612858] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.224 [2024-09-30 14:11:37.642718] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:33.224 [2024-09-30 14:11:37.642830] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.224 [2024-09-30 14:11:37.642868] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.224 [2024-09-30 14:11:37.642875] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:33.224 "name": "raid_bdev1", 00:12:33.224 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:33.224 "strip_size_kb": 0, 00:12:33.224 "state": "online", 00:12:33.224 "raid_level": "raid1", 00:12:33.224 "superblock": true, 00:12:33.224 "num_base_bdevs": 4, 00:12:33.224 "num_base_bdevs_discovered": 2, 00:12:33.224 "num_base_bdevs_operational": 2, 00:12:33.224 "base_bdevs_list": [ 00:12:33.224 { 00:12:33.224 "name": null, 00:12:33.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.224 "is_configured": false, 00:12:33.224 "data_offset": 0, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": null, 00:12:33.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.224 "is_configured": false, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": "BaseBdev3", 00:12:33.224 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:33.224 "is_configured": true, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 }, 00:12:33.224 { 00:12:33.224 "name": "BaseBdev4", 00:12:33.224 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:33.224 "is_configured": true, 00:12:33.224 "data_offset": 2048, 00:12:33.224 "data_size": 63488 00:12:33.224 } 00:12:33.224 ] 00:12:33.224 }' 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:33.224 14:11:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.484 14:11:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:33.484 14:11:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.484 14:11:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.484 [2024-09-30 14:11:38.074098] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:33.484 [2024-09-30 14:11:38.074229] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:33.484 [2024-09-30 14:11:38.074274] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:12:33.484 [2024-09-30 14:11:38.074303] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:33.484 [2024-09-30 14:11:38.074768] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:33.484 [2024-09-30 14:11:38.074828] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:33.484 [2024-09-30 14:11:38.074948] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:33.484 [2024-09-30 14:11:38.074985] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:33.484 [2024-09-30 14:11:38.075055] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:33.484 [2024-09-30 14:11:38.075123] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:33.484 [2024-09-30 14:11:38.078399] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ef0 00:12:33.484 spare 00:12:33.484 14:11:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.484 [2024-09-30 14:11:38.080278] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:33.484 14:11:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.866 "name": "raid_bdev1", 00:12:34.866 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:34.866 "strip_size_kb": 0, 00:12:34.866 "state": "online", 00:12:34.866 "raid_level": "raid1", 00:12:34.866 "superblock": true, 00:12:34.866 "num_base_bdevs": 4, 00:12:34.866 "num_base_bdevs_discovered": 3, 00:12:34.866 "num_base_bdevs_operational": 3, 00:12:34.866 "process": { 00:12:34.866 "type": "rebuild", 00:12:34.866 "target": "spare", 00:12:34.866 "progress": { 00:12:34.866 "blocks": 20480, 00:12:34.866 "percent": 32 00:12:34.866 } 00:12:34.866 }, 00:12:34.866 "base_bdevs_list": [ 00:12:34.866 { 00:12:34.866 "name": "spare", 00:12:34.866 "uuid": "da49214d-45ab-5291-82f9-e56bd2a80ce7", 00:12:34.866 "is_configured": true, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": null, 00:12:34.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.866 "is_configured": false, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": "BaseBdev3", 00:12:34.866 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:34.866 "is_configured": true, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": "BaseBdev4", 00:12:34.866 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:34.866 "is_configured": true, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 } 00:12:34.866 ] 00:12:34.866 }' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.866 [2024-09-30 14:11:39.236991] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:34.866 [2024-09-30 14:11:39.284842] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:34.866 [2024-09-30 14:11:39.284970] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:34.866 [2024-09-30 14:11:39.285006] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:34.866 [2024-09-30 14:11:39.285029] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.866 "name": "raid_bdev1", 00:12:34.866 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:34.866 "strip_size_kb": 0, 00:12:34.866 "state": "online", 00:12:34.866 "raid_level": "raid1", 00:12:34.866 "superblock": true, 00:12:34.866 "num_base_bdevs": 4, 00:12:34.866 "num_base_bdevs_discovered": 2, 00:12:34.866 "num_base_bdevs_operational": 2, 00:12:34.866 "base_bdevs_list": [ 00:12:34.866 { 00:12:34.866 "name": null, 00:12:34.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.866 "is_configured": false, 00:12:34.866 "data_offset": 0, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": null, 00:12:34.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.866 "is_configured": false, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": "BaseBdev3", 00:12:34.866 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:34.866 "is_configured": true, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 }, 00:12:34.866 { 00:12:34.866 "name": "BaseBdev4", 00:12:34.866 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:34.866 "is_configured": true, 00:12:34.866 "data_offset": 2048, 00:12:34.866 "data_size": 63488 00:12:34.866 } 00:12:34.866 ] 00:12:34.866 }' 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.866 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.126 "name": "raid_bdev1", 00:12:35.126 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:35.126 "strip_size_kb": 0, 00:12:35.126 "state": "online", 00:12:35.126 "raid_level": "raid1", 00:12:35.126 "superblock": true, 00:12:35.126 "num_base_bdevs": 4, 00:12:35.126 "num_base_bdevs_discovered": 2, 00:12:35.126 "num_base_bdevs_operational": 2, 00:12:35.126 "base_bdevs_list": [ 00:12:35.126 { 00:12:35.126 "name": null, 00:12:35.126 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.126 "is_configured": false, 00:12:35.126 "data_offset": 0, 00:12:35.126 "data_size": 63488 00:12:35.126 }, 00:12:35.126 { 00:12:35.126 "name": null, 00:12:35.126 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.126 "is_configured": false, 00:12:35.126 "data_offset": 2048, 00:12:35.126 "data_size": 63488 00:12:35.126 }, 00:12:35.126 { 00:12:35.126 "name": "BaseBdev3", 00:12:35.126 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:35.126 "is_configured": true, 00:12:35.126 "data_offset": 2048, 00:12:35.126 "data_size": 63488 00:12:35.126 }, 00:12:35.126 { 00:12:35.126 "name": "BaseBdev4", 00:12:35.126 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:35.126 "is_configured": true, 00:12:35.126 "data_offset": 2048, 00:12:35.126 "data_size": 63488 00:12:35.126 } 00:12:35.126 ] 00:12:35.126 }' 00:12:35.126 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.386 [2024-09-30 14:11:39.835974] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:35.386 [2024-09-30 14:11:39.836082] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:35.386 [2024-09-30 14:11:39.836134] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:12:35.386 [2024-09-30 14:11:39.836164] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:35.386 [2024-09-30 14:11:39.836596] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:35.386 [2024-09-30 14:11:39.836657] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:35.386 [2024-09-30 14:11:39.836749] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:35.386 [2024-09-30 14:11:39.836799] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:35.386 [2024-09-30 14:11:39.836838] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:35.386 [2024-09-30 14:11:39.836882] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:35.386 BaseBdev1 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.386 14:11:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.324 "name": "raid_bdev1", 00:12:36.324 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:36.324 "strip_size_kb": 0, 00:12:36.324 "state": "online", 00:12:36.324 "raid_level": "raid1", 00:12:36.324 "superblock": true, 00:12:36.324 "num_base_bdevs": 4, 00:12:36.324 "num_base_bdevs_discovered": 2, 00:12:36.324 "num_base_bdevs_operational": 2, 00:12:36.324 "base_bdevs_list": [ 00:12:36.324 { 00:12:36.324 "name": null, 00:12:36.324 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.324 "is_configured": false, 00:12:36.324 "data_offset": 0, 00:12:36.324 "data_size": 63488 00:12:36.324 }, 00:12:36.324 { 00:12:36.324 "name": null, 00:12:36.324 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.324 "is_configured": false, 00:12:36.324 "data_offset": 2048, 00:12:36.324 "data_size": 63488 00:12:36.324 }, 00:12:36.324 { 00:12:36.324 "name": "BaseBdev3", 00:12:36.324 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:36.324 "is_configured": true, 00:12:36.324 "data_offset": 2048, 00:12:36.324 "data_size": 63488 00:12:36.324 }, 00:12:36.324 { 00:12:36.324 "name": "BaseBdev4", 00:12:36.324 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:36.324 "is_configured": true, 00:12:36.324 "data_offset": 2048, 00:12:36.324 "data_size": 63488 00:12:36.324 } 00:12:36.324 ] 00:12:36.324 }' 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.324 14:11:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:36.893 "name": "raid_bdev1", 00:12:36.893 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:36.893 "strip_size_kb": 0, 00:12:36.893 "state": "online", 00:12:36.893 "raid_level": "raid1", 00:12:36.893 "superblock": true, 00:12:36.893 "num_base_bdevs": 4, 00:12:36.893 "num_base_bdevs_discovered": 2, 00:12:36.893 "num_base_bdevs_operational": 2, 00:12:36.893 "base_bdevs_list": [ 00:12:36.893 { 00:12:36.893 "name": null, 00:12:36.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.893 "is_configured": false, 00:12:36.893 "data_offset": 0, 00:12:36.893 "data_size": 63488 00:12:36.893 }, 00:12:36.893 { 00:12:36.893 "name": null, 00:12:36.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.893 "is_configured": false, 00:12:36.893 "data_offset": 2048, 00:12:36.893 "data_size": 63488 00:12:36.893 }, 00:12:36.893 { 00:12:36.893 "name": "BaseBdev3", 00:12:36.893 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:36.893 "is_configured": true, 00:12:36.893 "data_offset": 2048, 00:12:36.893 "data_size": 63488 00:12:36.893 }, 00:12:36.893 { 00:12:36.893 "name": "BaseBdev4", 00:12:36.893 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:36.893 "is_configured": true, 00:12:36.893 "data_offset": 2048, 00:12:36.893 "data_size": 63488 00:12:36.893 } 00:12:36.893 ] 00:12:36.893 }' 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.893 [2024-09-30 14:11:41.405286] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:36.893 [2024-09-30 14:11:41.405513] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:36.893 [2024-09-30 14:11:41.405565] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:36.893 request: 00:12:36.893 { 00:12:36.893 "base_bdev": "BaseBdev1", 00:12:36.893 "raid_bdev": "raid_bdev1", 00:12:36.893 "method": "bdev_raid_add_base_bdev", 00:12:36.893 "req_id": 1 00:12:36.893 } 00:12:36.893 Got JSON-RPC error response 00:12:36.893 response: 00:12:36.893 { 00:12:36.893 "code": -22, 00:12:36.893 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:36.893 } 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:36.893 14:11:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:37.831 "name": "raid_bdev1", 00:12:37.831 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:37.831 "strip_size_kb": 0, 00:12:37.831 "state": "online", 00:12:37.831 "raid_level": "raid1", 00:12:37.831 "superblock": true, 00:12:37.831 "num_base_bdevs": 4, 00:12:37.831 "num_base_bdevs_discovered": 2, 00:12:37.831 "num_base_bdevs_operational": 2, 00:12:37.831 "base_bdevs_list": [ 00:12:37.831 { 00:12:37.831 "name": null, 00:12:37.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.831 "is_configured": false, 00:12:37.831 "data_offset": 0, 00:12:37.831 "data_size": 63488 00:12:37.831 }, 00:12:37.831 { 00:12:37.831 "name": null, 00:12:37.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.831 "is_configured": false, 00:12:37.831 "data_offset": 2048, 00:12:37.831 "data_size": 63488 00:12:37.831 }, 00:12:37.831 { 00:12:37.831 "name": "BaseBdev3", 00:12:37.831 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:37.831 "is_configured": true, 00:12:37.831 "data_offset": 2048, 00:12:37.831 "data_size": 63488 00:12:37.831 }, 00:12:37.831 { 00:12:37.831 "name": "BaseBdev4", 00:12:37.831 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:37.831 "is_configured": true, 00:12:37.831 "data_offset": 2048, 00:12:37.831 "data_size": 63488 00:12:37.831 } 00:12:37.831 ] 00:12:37.831 }' 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:37.831 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.400 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.400 "name": "raid_bdev1", 00:12:38.400 "uuid": "7b48c030-f080-45d0-b8a9-42a8c28cf1f7", 00:12:38.400 "strip_size_kb": 0, 00:12:38.400 "state": "online", 00:12:38.400 "raid_level": "raid1", 00:12:38.400 "superblock": true, 00:12:38.400 "num_base_bdevs": 4, 00:12:38.400 "num_base_bdevs_discovered": 2, 00:12:38.400 "num_base_bdevs_operational": 2, 00:12:38.400 "base_bdevs_list": [ 00:12:38.400 { 00:12:38.400 "name": null, 00:12:38.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.401 "is_configured": false, 00:12:38.401 "data_offset": 0, 00:12:38.401 "data_size": 63488 00:12:38.401 }, 00:12:38.401 { 00:12:38.401 "name": null, 00:12:38.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.401 "is_configured": false, 00:12:38.401 "data_offset": 2048, 00:12:38.401 "data_size": 63488 00:12:38.401 }, 00:12:38.401 { 00:12:38.401 "name": "BaseBdev3", 00:12:38.401 "uuid": "766cb61c-5de5-5cf6-a061-cb1a6de54ac4", 00:12:38.401 "is_configured": true, 00:12:38.401 "data_offset": 2048, 00:12:38.401 "data_size": 63488 00:12:38.401 }, 00:12:38.401 { 00:12:38.401 "name": "BaseBdev4", 00:12:38.401 "uuid": "3781197c-9069-5cf7-810c-cbf5c479325e", 00:12:38.401 "is_configured": true, 00:12:38.401 "data_offset": 2048, 00:12:38.401 "data_size": 63488 00:12:38.401 } 00:12:38.401 ] 00:12:38.401 }' 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 89640 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 89640 ']' 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 89640 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:38.401 14:11:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89640 00:12:38.401 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:38.401 killing process with pid 89640 00:12:38.401 Received shutdown signal, test time was about 60.000000 seconds 00:12:38.401 00:12:38.401 Latency(us) 00:12:38.401 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:38.401 =================================================================================================================== 00:12:38.401 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:38.401 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:38.401 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89640' 00:12:38.401 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 89640 00:12:38.401 [2024-09-30 14:11:43.014261] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:38.401 [2024-09-30 14:11:43.014371] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:38.401 [2024-09-30 14:11:43.014454] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:38.401 [2024-09-30 14:11:43.014465] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:12:38.401 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 89640 00:12:38.661 [2024-09-30 14:11:43.062741] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:38.661 14:11:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:38.661 00:12:38.661 real 0m22.832s 00:12:38.661 user 0m27.733s 00:12:38.661 sys 0m3.644s 00:12:38.661 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:38.661 ************************************ 00:12:38.661 END TEST raid_rebuild_test_sb 00:12:38.661 ************************************ 00:12:38.661 14:11:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.920 14:11:43 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:38.920 14:11:43 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:38.920 14:11:43 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:38.920 14:11:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:38.920 ************************************ 00:12:38.920 START TEST raid_rebuild_test_io 00:12:38.920 ************************************ 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false true true 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.920 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=90376 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 90376 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 90376 ']' 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:38.921 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:38.921 14:11:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.921 [2024-09-30 14:11:43.467451] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:12:38.921 [2024-09-30 14:11:43.467666] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --matchI/O size of 3145728 is greater than zero copy threshold (65536). 00:12:38.921 Zero copy mechanism will not be used. 00:12:38.921 -allocations --file-prefix=spdk_pid90376 ] 00:12:39.180 [2024-09-30 14:11:43.598306] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:12:39.180 [2024-09-30 14:11:43.627526] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:39.180 [2024-09-30 14:11:43.671090] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:39.180 [2024-09-30 14:11:43.711679] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:39.180 [2024-09-30 14:11:43.711718] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.748 BaseBdev1_malloc 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:39.748 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.749 [2024-09-30 14:11:44.313167] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:39.749 [2024-09-30 14:11:44.313293] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.749 [2024-09-30 14:11:44.313337] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:39.749 [2024-09-30 14:11:44.313382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.749 [2024-09-30 14:11:44.315427] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.749 [2024-09-30 14:11:44.315520] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:39.749 BaseBdev1 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.749 BaseBdev2_malloc 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.749 [2024-09-30 14:11:44.355430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:39.749 [2024-09-30 14:11:44.355652] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.749 [2024-09-30 14:11:44.355742] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:39.749 [2024-09-30 14:11:44.355824] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.749 BaseBdev2 00:12:39.749 [2024-09-30 14:11:44.360653] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.749 [2024-09-30 14:11:44.360727] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.749 BaseBdev3_malloc 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.749 [2024-09-30 14:11:44.386744] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:39.749 [2024-09-30 14:11:44.386845] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.749 [2024-09-30 14:11:44.386881] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:39.749 [2024-09-30 14:11:44.386912] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.749 [2024-09-30 14:11:44.388970] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.749 [2024-09-30 14:11:44.389041] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:39.749 BaseBdev3 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.749 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 BaseBdev4_malloc 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 [2024-09-30 14:11:44.416050] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:40.014 [2024-09-30 14:11:44.416158] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:40.014 [2024-09-30 14:11:44.416194] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:40.014 [2024-09-30 14:11:44.416225] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:40.014 [2024-09-30 14:11:44.418265] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:40.014 [2024-09-30 14:11:44.418336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:40.014 BaseBdev4 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 spare_malloc 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 spare_delay 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 [2024-09-30 14:11:44.456628] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:40.014 [2024-09-30 14:11:44.456735] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:40.014 [2024-09-30 14:11:44.456783] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:40.014 [2024-09-30 14:11:44.456815] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:40.014 [2024-09-30 14:11:44.458859] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:40.014 [2024-09-30 14:11:44.458925] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:40.014 spare 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.014 [2024-09-30 14:11:44.468679] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:40.014 [2024-09-30 14:11:44.470541] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:40.014 [2024-09-30 14:11:44.470635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:40.014 [2024-09-30 14:11:44.470698] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:40.014 [2024-09-30 14:11:44.470814] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:40.014 [2024-09-30 14:11:44.470861] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:40.014 [2024-09-30 14:11:44.471130] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:40.014 [2024-09-30 14:11:44.471298] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:40.014 [2024-09-30 14:11:44.471336] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:40.014 [2024-09-30 14:11:44.471499] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:40.014 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.015 "name": "raid_bdev1", 00:12:40.015 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:40.015 "strip_size_kb": 0, 00:12:40.015 "state": "online", 00:12:40.015 "raid_level": "raid1", 00:12:40.015 "superblock": false, 00:12:40.015 "num_base_bdevs": 4, 00:12:40.015 "num_base_bdevs_discovered": 4, 00:12:40.015 "num_base_bdevs_operational": 4, 00:12:40.015 "base_bdevs_list": [ 00:12:40.015 { 00:12:40.015 "name": "BaseBdev1", 00:12:40.015 "uuid": "9e8d501d-4c07-5418-9eab-44b119ffdc11", 00:12:40.015 "is_configured": true, 00:12:40.015 "data_offset": 0, 00:12:40.015 "data_size": 65536 00:12:40.015 }, 00:12:40.015 { 00:12:40.015 "name": "BaseBdev2", 00:12:40.015 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:40.015 "is_configured": true, 00:12:40.015 "data_offset": 0, 00:12:40.015 "data_size": 65536 00:12:40.015 }, 00:12:40.015 { 00:12:40.015 "name": "BaseBdev3", 00:12:40.015 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:40.015 "is_configured": true, 00:12:40.015 "data_offset": 0, 00:12:40.015 "data_size": 65536 00:12:40.015 }, 00:12:40.015 { 00:12:40.015 "name": "BaseBdev4", 00:12:40.015 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:40.015 "is_configured": true, 00:12:40.015 "data_offset": 0, 00:12:40.015 "data_size": 65536 00:12:40.015 } 00:12:40.015 ] 00:12:40.015 }' 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.015 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.274 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:40.274 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:40.274 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.274 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.534 [2024-09-30 14:11:44.932273] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.534 14:11:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.534 [2024-09-30 14:11:45.019748] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.534 "name": "raid_bdev1", 00:12:40.534 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:40.534 "strip_size_kb": 0, 00:12:40.534 "state": "online", 00:12:40.534 "raid_level": "raid1", 00:12:40.534 "superblock": false, 00:12:40.534 "num_base_bdevs": 4, 00:12:40.534 "num_base_bdevs_discovered": 3, 00:12:40.534 "num_base_bdevs_operational": 3, 00:12:40.534 "base_bdevs_list": [ 00:12:40.534 { 00:12:40.534 "name": null, 00:12:40.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.534 "is_configured": false, 00:12:40.534 "data_offset": 0, 00:12:40.534 "data_size": 65536 00:12:40.534 }, 00:12:40.534 { 00:12:40.534 "name": "BaseBdev2", 00:12:40.534 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:40.534 "is_configured": true, 00:12:40.534 "data_offset": 0, 00:12:40.534 "data_size": 65536 00:12:40.534 }, 00:12:40.534 { 00:12:40.534 "name": "BaseBdev3", 00:12:40.534 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:40.534 "is_configured": true, 00:12:40.534 "data_offset": 0, 00:12:40.534 "data_size": 65536 00:12:40.534 }, 00:12:40.534 { 00:12:40.534 "name": "BaseBdev4", 00:12:40.534 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:40.534 "is_configured": true, 00:12:40.534 "data_offset": 0, 00:12:40.534 "data_size": 65536 00:12:40.534 } 00:12:40.534 ] 00:12:40.534 }' 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.534 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.534 [2024-09-30 14:11:45.109628] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:12:40.534 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:40.534 Zero copy mechanism will not be used. 00:12:40.534 Running I/O for 60 seconds... 00:12:40.794 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:40.794 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.794 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.794 [2024-09-30 14:11:45.440815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:41.054 14:11:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:41.054 14:11:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:41.054 [2024-09-30 14:11:45.488224] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:12:41.054 [2024-09-30 14:11:45.490174] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:41.054 [2024-09-30 14:11:45.591979] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:41.054 [2024-09-30 14:11:45.592350] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:41.314 [2024-09-30 14:11:45.808969] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:41.314 [2024-09-30 14:11:45.809723] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:41.573 165.00 IOPS, 495.00 MiB/s [2024-09-30 14:11:46.130672] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:41.833 [2024-09-30 14:11:46.265474] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:41.833 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:42.093 "name": "raid_bdev1", 00:12:42.093 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:42.093 "strip_size_kb": 0, 00:12:42.093 "state": "online", 00:12:42.093 "raid_level": "raid1", 00:12:42.093 "superblock": false, 00:12:42.093 "num_base_bdevs": 4, 00:12:42.093 "num_base_bdevs_discovered": 4, 00:12:42.093 "num_base_bdevs_operational": 4, 00:12:42.093 "process": { 00:12:42.093 "type": "rebuild", 00:12:42.093 "target": "spare", 00:12:42.093 "progress": { 00:12:42.093 "blocks": 12288, 00:12:42.093 "percent": 18 00:12:42.093 } 00:12:42.093 }, 00:12:42.093 "base_bdevs_list": [ 00:12:42.093 { 00:12:42.093 "name": "spare", 00:12:42.093 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:42.093 "is_configured": true, 00:12:42.093 "data_offset": 0, 00:12:42.093 "data_size": 65536 00:12:42.093 }, 00:12:42.093 { 00:12:42.093 "name": "BaseBdev2", 00:12:42.093 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:42.093 "is_configured": true, 00:12:42.093 "data_offset": 0, 00:12:42.093 "data_size": 65536 00:12:42.093 }, 00:12:42.093 { 00:12:42.093 "name": "BaseBdev3", 00:12:42.093 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:42.093 "is_configured": true, 00:12:42.093 "data_offset": 0, 00:12:42.093 "data_size": 65536 00:12:42.093 }, 00:12:42.093 { 00:12:42.093 "name": "BaseBdev4", 00:12:42.093 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:42.093 "is_configured": true, 00:12:42.093 "data_offset": 0, 00:12:42.093 "data_size": 65536 00:12:42.093 } 00:12:42.093 ] 00:12:42.093 }' 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:42.093 [2024-09-30 14:11:46.591353] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:42.093 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.093 [2024-09-30 14:11:46.599375] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:42.093 [2024-09-30 14:11:46.723551] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:42.093 [2024-09-30 14:11:46.726812] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:42.093 [2024-09-30 14:11:46.726853] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:42.093 [2024-09-30 14:11:46.726864] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:42.093 [2024-09-30 14:11:46.748873] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000062f0 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:42.354 "name": "raid_bdev1", 00:12:42.354 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:42.354 "strip_size_kb": 0, 00:12:42.354 "state": "online", 00:12:42.354 "raid_level": "raid1", 00:12:42.354 "superblock": false, 00:12:42.354 "num_base_bdevs": 4, 00:12:42.354 "num_base_bdevs_discovered": 3, 00:12:42.354 "num_base_bdevs_operational": 3, 00:12:42.354 "base_bdevs_list": [ 00:12:42.354 { 00:12:42.354 "name": null, 00:12:42.354 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.354 "is_configured": false, 00:12:42.354 "data_offset": 0, 00:12:42.354 "data_size": 65536 00:12:42.354 }, 00:12:42.354 { 00:12:42.354 "name": "BaseBdev2", 00:12:42.354 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:42.354 "is_configured": true, 00:12:42.354 "data_offset": 0, 00:12:42.354 "data_size": 65536 00:12:42.354 }, 00:12:42.354 { 00:12:42.354 "name": "BaseBdev3", 00:12:42.354 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:42.354 "is_configured": true, 00:12:42.354 "data_offset": 0, 00:12:42.354 "data_size": 65536 00:12:42.354 }, 00:12:42.354 { 00:12:42.354 "name": "BaseBdev4", 00:12:42.354 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:42.354 "is_configured": true, 00:12:42.354 "data_offset": 0, 00:12:42.354 "data_size": 65536 00:12:42.354 } 00:12:42.354 ] 00:12:42.354 }' 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:42.354 14:11:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.614 152.00 IOPS, 456.00 MiB/s 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.614 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:42.615 "name": "raid_bdev1", 00:12:42.615 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:42.615 "strip_size_kb": 0, 00:12:42.615 "state": "online", 00:12:42.615 "raid_level": "raid1", 00:12:42.615 "superblock": false, 00:12:42.615 "num_base_bdevs": 4, 00:12:42.615 "num_base_bdevs_discovered": 3, 00:12:42.615 "num_base_bdevs_operational": 3, 00:12:42.615 "base_bdevs_list": [ 00:12:42.615 { 00:12:42.615 "name": null, 00:12:42.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.615 "is_configured": false, 00:12:42.615 "data_offset": 0, 00:12:42.615 "data_size": 65536 00:12:42.615 }, 00:12:42.615 { 00:12:42.615 "name": "BaseBdev2", 00:12:42.615 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:42.615 "is_configured": true, 00:12:42.615 "data_offset": 0, 00:12:42.615 "data_size": 65536 00:12:42.615 }, 00:12:42.615 { 00:12:42.615 "name": "BaseBdev3", 00:12:42.615 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:42.615 "is_configured": true, 00:12:42.615 "data_offset": 0, 00:12:42.615 "data_size": 65536 00:12:42.615 }, 00:12:42.615 { 00:12:42.615 "name": "BaseBdev4", 00:12:42.615 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:42.615 "is_configured": true, 00:12:42.615 "data_offset": 0, 00:12:42.615 "data_size": 65536 00:12:42.615 } 00:12:42.615 ] 00:12:42.615 }' 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:42.615 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.883 [2024-09-30 14:11:47.292870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:42.883 14:11:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:42.883 [2024-09-30 14:11:47.345516] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:12:42.883 [2024-09-30 14:11:47.347436] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:42.883 [2024-09-30 14:11:47.461469] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:42.883 [2024-09-30 14:11:47.462636] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:43.143 [2024-09-30 14:11:47.679944] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:43.143 [2024-09-30 14:11:47.680126] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:43.402 [2024-09-30 14:11:48.051846] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:43.921 146.67 IOPS, 440.00 MiB/s 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:43.921 "name": "raid_bdev1", 00:12:43.921 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:43.921 "strip_size_kb": 0, 00:12:43.921 "state": "online", 00:12:43.921 "raid_level": "raid1", 00:12:43.921 "superblock": false, 00:12:43.921 "num_base_bdevs": 4, 00:12:43.921 "num_base_bdevs_discovered": 4, 00:12:43.921 "num_base_bdevs_operational": 4, 00:12:43.921 "process": { 00:12:43.921 "type": "rebuild", 00:12:43.921 "target": "spare", 00:12:43.921 "progress": { 00:12:43.921 "blocks": 12288, 00:12:43.921 "percent": 18 00:12:43.921 } 00:12:43.921 }, 00:12:43.921 "base_bdevs_list": [ 00:12:43.921 { 00:12:43.921 "name": "spare", 00:12:43.921 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:43.921 "is_configured": true, 00:12:43.921 "data_offset": 0, 00:12:43.921 "data_size": 65536 00:12:43.921 }, 00:12:43.921 { 00:12:43.921 "name": "BaseBdev2", 00:12:43.921 "uuid": "b89e45ba-9eae-569e-9753-4ea8452ca03d", 00:12:43.921 "is_configured": true, 00:12:43.921 "data_offset": 0, 00:12:43.921 "data_size": 65536 00:12:43.921 }, 00:12:43.921 { 00:12:43.921 "name": "BaseBdev3", 00:12:43.921 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:43.921 "is_configured": true, 00:12:43.921 "data_offset": 0, 00:12:43.921 "data_size": 65536 00:12:43.921 }, 00:12:43.921 { 00:12:43.921 "name": "BaseBdev4", 00:12:43.921 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:43.921 "is_configured": true, 00:12:43.921 "data_offset": 0, 00:12:43.921 "data_size": 65536 00:12:43.921 } 00:12:43.921 ] 00:12:43.921 }' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:43.921 [2024-09-30 14:11:48.389136] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:43.921 [2024-09-30 14:11:48.390318] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:43.921 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.921 [2024-09-30 14:11:48.481191] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:44.181 [2024-09-30 14:11:48.597872] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:44.181 [2024-09-30 14:11:48.598087] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:44.181 [2024-09-30 14:11:48.705370] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000062f0 00:12:44.181 [2024-09-30 14:11:48.705436] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006490 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.181 "name": "raid_bdev1", 00:12:44.181 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:44.181 "strip_size_kb": 0, 00:12:44.181 "state": "online", 00:12:44.181 "raid_level": "raid1", 00:12:44.181 "superblock": false, 00:12:44.181 "num_base_bdevs": 4, 00:12:44.181 "num_base_bdevs_discovered": 3, 00:12:44.181 "num_base_bdevs_operational": 3, 00:12:44.181 "process": { 00:12:44.181 "type": "rebuild", 00:12:44.181 "target": "spare", 00:12:44.181 "progress": { 00:12:44.181 "blocks": 16384, 00:12:44.181 "percent": 25 00:12:44.181 } 00:12:44.181 }, 00:12:44.181 "base_bdevs_list": [ 00:12:44.181 { 00:12:44.181 "name": "spare", 00:12:44.181 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:44.181 "is_configured": true, 00:12:44.181 "data_offset": 0, 00:12:44.181 "data_size": 65536 00:12:44.181 }, 00:12:44.181 { 00:12:44.181 "name": null, 00:12:44.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.181 "is_configured": false, 00:12:44.181 "data_offset": 0, 00:12:44.181 "data_size": 65536 00:12:44.181 }, 00:12:44.181 { 00:12:44.181 "name": "BaseBdev3", 00:12:44.181 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:44.181 "is_configured": true, 00:12:44.181 "data_offset": 0, 00:12:44.181 "data_size": 65536 00:12:44.181 }, 00:12:44.181 { 00:12:44.181 "name": "BaseBdev4", 00:12:44.181 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:44.181 "is_configured": true, 00:12:44.181 "data_offset": 0, 00:12:44.181 "data_size": 65536 00:12:44.181 } 00:12:44.181 ] 00:12:44.181 }' 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:44.181 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=389 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.441 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.441 "name": "raid_bdev1", 00:12:44.441 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:44.441 "strip_size_kb": 0, 00:12:44.441 "state": "online", 00:12:44.441 "raid_level": "raid1", 00:12:44.441 "superblock": false, 00:12:44.441 "num_base_bdevs": 4, 00:12:44.441 "num_base_bdevs_discovered": 3, 00:12:44.441 "num_base_bdevs_operational": 3, 00:12:44.441 "process": { 00:12:44.441 "type": "rebuild", 00:12:44.441 "target": "spare", 00:12:44.441 "progress": { 00:12:44.441 "blocks": 18432, 00:12:44.441 "percent": 28 00:12:44.441 } 00:12:44.441 }, 00:12:44.441 "base_bdevs_list": [ 00:12:44.441 { 00:12:44.441 "name": "spare", 00:12:44.441 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:44.441 "is_configured": true, 00:12:44.441 "data_offset": 0, 00:12:44.442 "data_size": 65536 00:12:44.442 }, 00:12:44.442 { 00:12:44.442 "name": null, 00:12:44.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.442 "is_configured": false, 00:12:44.442 "data_offset": 0, 00:12:44.442 "data_size": 65536 00:12:44.442 }, 00:12:44.442 { 00:12:44.442 "name": "BaseBdev3", 00:12:44.442 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:44.442 "is_configured": true, 00:12:44.442 "data_offset": 0, 00:12:44.442 "data_size": 65536 00:12:44.442 }, 00:12:44.442 { 00:12:44.442 "name": "BaseBdev4", 00:12:44.442 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:44.442 "is_configured": true, 00:12:44.442 "data_offset": 0, 00:12:44.442 "data_size": 65536 00:12:44.442 } 00:12:44.442 ] 00:12:44.442 }' 00:12:44.442 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.442 [2024-09-30 14:11:48.935169] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:44.442 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:44.442 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.442 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:44.442 14:11:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:44.702 117.75 IOPS, 353.25 MiB/s [2024-09-30 14:11:49.159778] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:44.961 [2024-09-30 14:11:49.378279] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:44.961 [2024-09-30 14:11:49.599708] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:45.530 14:11:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.530 "name": "raid_bdev1", 00:12:45.530 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:45.530 "strip_size_kb": 0, 00:12:45.530 "state": "online", 00:12:45.530 "raid_level": "raid1", 00:12:45.530 "superblock": false, 00:12:45.530 "num_base_bdevs": 4, 00:12:45.530 "num_base_bdevs_discovered": 3, 00:12:45.530 "num_base_bdevs_operational": 3, 00:12:45.530 "process": { 00:12:45.530 "type": "rebuild", 00:12:45.530 "target": "spare", 00:12:45.530 "progress": { 00:12:45.530 "blocks": 34816, 00:12:45.530 "percent": 53 00:12:45.530 } 00:12:45.530 }, 00:12:45.530 "base_bdevs_list": [ 00:12:45.530 { 00:12:45.530 "name": "spare", 00:12:45.530 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:45.530 "is_configured": true, 00:12:45.530 "data_offset": 0, 00:12:45.530 "data_size": 65536 00:12:45.530 }, 00:12:45.530 { 00:12:45.530 "name": null, 00:12:45.530 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.530 "is_configured": false, 00:12:45.530 "data_offset": 0, 00:12:45.530 "data_size": 65536 00:12:45.530 }, 00:12:45.530 { 00:12:45.530 "name": "BaseBdev3", 00:12:45.530 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:45.530 "is_configured": true, 00:12:45.530 "data_offset": 0, 00:12:45.530 "data_size": 65536 00:12:45.530 }, 00:12:45.530 { 00:12:45.530 "name": "BaseBdev4", 00:12:45.530 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:45.530 "is_configured": true, 00:12:45.530 "data_offset": 0, 00:12:45.530 "data_size": 65536 00:12:45.530 } 00:12:45.530 ] 00:12:45.530 }' 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.530 105.60 IOPS, 316.80 MiB/s 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:45.530 14:11:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:45.530 [2024-09-30 14:11:50.183933] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:46.099 [2024-09-30 14:11:50.619983] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:46.358 [2024-09-30 14:11:50.934398] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:46.618 95.50 IOPS, 286.50 MiB/s 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.618 [2024-09-30 14:11:51.177401] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.618 "name": "raid_bdev1", 00:12:46.618 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:46.618 "strip_size_kb": 0, 00:12:46.618 "state": "online", 00:12:46.618 "raid_level": "raid1", 00:12:46.618 "superblock": false, 00:12:46.618 "num_base_bdevs": 4, 00:12:46.618 "num_base_bdevs_discovered": 3, 00:12:46.618 "num_base_bdevs_operational": 3, 00:12:46.618 "process": { 00:12:46.618 "type": "rebuild", 00:12:46.618 "target": "spare", 00:12:46.618 "progress": { 00:12:46.618 "blocks": 55296, 00:12:46.618 "percent": 84 00:12:46.618 } 00:12:46.618 }, 00:12:46.618 "base_bdevs_list": [ 00:12:46.618 { 00:12:46.618 "name": "spare", 00:12:46.618 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:46.618 "is_configured": true, 00:12:46.618 "data_offset": 0, 00:12:46.618 "data_size": 65536 00:12:46.618 }, 00:12:46.618 { 00:12:46.618 "name": null, 00:12:46.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.618 "is_configured": false, 00:12:46.618 "data_offset": 0, 00:12:46.618 "data_size": 65536 00:12:46.618 }, 00:12:46.618 { 00:12:46.618 "name": "BaseBdev3", 00:12:46.618 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:46.618 "is_configured": true, 00:12:46.618 "data_offset": 0, 00:12:46.618 "data_size": 65536 00:12:46.618 }, 00:12:46.618 { 00:12:46.618 "name": "BaseBdev4", 00:12:46.618 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:46.618 "is_configured": true, 00:12:46.618 "data_offset": 0, 00:12:46.618 "data_size": 65536 00:12:46.618 } 00:12:46.618 ] 00:12:46.618 }' 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:46.618 14:11:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:47.196 [2024-09-30 14:11:51.614790] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:47.196 [2024-09-30 14:11:51.714617] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:47.196 [2024-09-30 14:11:51.715855] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:47.766 87.71 IOPS, 263.14 MiB/s 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.766 "name": "raid_bdev1", 00:12:47.766 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:47.766 "strip_size_kb": 0, 00:12:47.766 "state": "online", 00:12:47.766 "raid_level": "raid1", 00:12:47.766 "superblock": false, 00:12:47.766 "num_base_bdevs": 4, 00:12:47.766 "num_base_bdevs_discovered": 3, 00:12:47.766 "num_base_bdevs_operational": 3, 00:12:47.766 "base_bdevs_list": [ 00:12:47.766 { 00:12:47.766 "name": "spare", 00:12:47.766 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:47.766 "is_configured": true, 00:12:47.766 "data_offset": 0, 00:12:47.766 "data_size": 65536 00:12:47.766 }, 00:12:47.766 { 00:12:47.766 "name": null, 00:12:47.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.766 "is_configured": false, 00:12:47.766 "data_offset": 0, 00:12:47.766 "data_size": 65536 00:12:47.766 }, 00:12:47.766 { 00:12:47.766 "name": "BaseBdev3", 00:12:47.766 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:47.766 "is_configured": true, 00:12:47.766 "data_offset": 0, 00:12:47.766 "data_size": 65536 00:12:47.766 }, 00:12:47.766 { 00:12:47.766 "name": "BaseBdev4", 00:12:47.766 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:47.766 "is_configured": true, 00:12:47.766 "data_offset": 0, 00:12:47.766 "data_size": 65536 00:12:47.766 } 00:12:47.766 ] 00:12:47.766 }' 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.766 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.026 "name": "raid_bdev1", 00:12:48.026 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:48.026 "strip_size_kb": 0, 00:12:48.026 "state": "online", 00:12:48.026 "raid_level": "raid1", 00:12:48.026 "superblock": false, 00:12:48.026 "num_base_bdevs": 4, 00:12:48.026 "num_base_bdevs_discovered": 3, 00:12:48.026 "num_base_bdevs_operational": 3, 00:12:48.026 "base_bdevs_list": [ 00:12:48.026 { 00:12:48.026 "name": "spare", 00:12:48.026 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:48.026 "is_configured": true, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.026 "name": null, 00:12:48.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.026 "is_configured": false, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.026 "name": "BaseBdev3", 00:12:48.026 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:48.026 "is_configured": true, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.026 "name": "BaseBdev4", 00:12:48.026 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:48.026 "is_configured": true, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 } 00:12:48.026 ] 00:12:48.026 }' 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.026 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.026 "name": "raid_bdev1", 00:12:48.026 "uuid": "1175460d-7101-490c-ad0e-1f0f0bf4f5d0", 00:12:48.026 "strip_size_kb": 0, 00:12:48.026 "state": "online", 00:12:48.026 "raid_level": "raid1", 00:12:48.026 "superblock": false, 00:12:48.026 "num_base_bdevs": 4, 00:12:48.026 "num_base_bdevs_discovered": 3, 00:12:48.026 "num_base_bdevs_operational": 3, 00:12:48.026 "base_bdevs_list": [ 00:12:48.026 { 00:12:48.026 "name": "spare", 00:12:48.026 "uuid": "b0187826-1d97-5c53-b35a-7486a8c2f102", 00:12:48.026 "is_configured": true, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.026 "name": null, 00:12:48.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.026 "is_configured": false, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.026 "name": "BaseBdev3", 00:12:48.026 "uuid": "9b1fd866-e1c3-519c-960e-9c76db5863ee", 00:12:48.026 "is_configured": true, 00:12:48.026 "data_offset": 0, 00:12:48.026 "data_size": 65536 00:12:48.026 }, 00:12:48.026 { 00:12:48.027 "name": "BaseBdev4", 00:12:48.027 "uuid": "15dabdbd-44b1-55df-9ece-171892bf6180", 00:12:48.027 "is_configured": true, 00:12:48.027 "data_offset": 0, 00:12:48.027 "data_size": 65536 00:12:48.027 } 00:12:48.027 ] 00:12:48.027 }' 00:12:48.027 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.027 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.596 14:11:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:48.596 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.597 14:11:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.597 [2024-09-30 14:11:53.006905] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:48.597 [2024-09-30 14:11:53.007015] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:48.597 00:12:48.597 Latency(us) 00:12:48.597 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:48.597 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:48.597 raid_bdev1 : 8.01 79.90 239.69 0.00 0.00 17567.65 268.30 109436.53 00:12:48.597 =================================================================================================================== 00:12:48.597 Total : 79.90 239.69 0.00 0.00 17567.65 268.30 109436.53 00:12:48.597 [2024-09-30 14:11:53.110060] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:48.597 [2024-09-30 14:11:53.110163] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:48.597 [2024-09-30 14:11:53.110275] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:48.597 [2024-09-30 14:11:53.110324] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:48.597 { 00:12:48.597 "results": [ 00:12:48.597 { 00:12:48.597 "job": "raid_bdev1", 00:12:48.597 "core_mask": "0x1", 00:12:48.597 "workload": "randrw", 00:12:48.597 "percentage": 50, 00:12:48.597 "status": "finished", 00:12:48.597 "queue_depth": 2, 00:12:48.597 "io_size": 3145728, 00:12:48.597 "runtime": 8.01044, 00:12:48.597 "iops": 79.89573606443591, 00:12:48.597 "mibps": 239.68720819330773, 00:12:48.597 "io_failed": 0, 00:12:48.597 "io_timeout": 0, 00:12:48.597 "avg_latency_us": 17567.649257641922, 00:12:48.597 "min_latency_us": 268.29694323144105, 00:12:48.597 "max_latency_us": 109436.5344978166 00:12:48.597 } 00:12:48.597 ], 00:12:48.597 "core_count": 1 00:12:48.597 } 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:48.597 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:48.857 /dev/nbd0 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:48.857 1+0 records in 00:12:48.857 1+0 records out 00:12:48.857 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00038878 s, 10.5 MB/s 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:48.857 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:49.118 /dev/nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:49.118 1+0 records in 00:12:49.118 1+0 records out 00:12:49.118 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000532116 s, 7.7 MB/s 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:49.118 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:49.379 14:11:53 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:49.639 /dev/nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:49.639 1+0 records in 00:12:49.639 1+0 records out 00:12:49.639 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000420255 s, 9.7 MB/s 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:49.639 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:49.899 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 90376 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 90376 ']' 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 90376 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90376 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:50.160 killing process with pid 90376 00:12:50.160 Received shutdown signal, test time was about 9.646104 seconds 00:12:50.160 00:12:50.160 Latency(us) 00:12:50.160 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:50.160 =================================================================================================================== 00:12:50.160 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90376' 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 90376 00:12:50.160 [2024-09-30 14:11:54.739178] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:50.160 14:11:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 90376 00:12:50.160 [2024-09-30 14:11:54.782745] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:50.421 14:11:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:50.421 00:12:50.421 real 0m11.645s 00:12:50.421 user 0m14.952s 00:12:50.421 sys 0m1.843s 00:12:50.421 14:11:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:50.421 ************************************ 00:12:50.421 END TEST raid_rebuild_test_io 00:12:50.421 ************************************ 00:12:50.421 14:11:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.681 14:11:55 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:12:50.681 14:11:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:50.681 14:11:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:50.681 14:11:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:50.681 ************************************ 00:12:50.681 START TEST raid_rebuild_test_sb_io 00:12:50.681 ************************************ 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true true true 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:50.681 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=90768 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 90768 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 90768 ']' 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:50.682 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:50.682 14:11:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.682 [2024-09-30 14:11:55.204598] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:12:50.682 [2024-09-30 14:11:55.204780] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --matchI/O size of 3145728 is greater than zero copy threshold (65536). 00:12:50.682 Zero copy mechanism will not be used. 00:12:50.682 -allocations --file-prefix=spdk_pid90768 ] 00:12:50.942 [2024-09-30 14:11:55.341633] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:12:50.942 [2024-09-30 14:11:55.369932] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:50.942 [2024-09-30 14:11:55.414805] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:50.942 [2024-09-30 14:11:55.457094] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:50.942 [2024-09-30 14:11:55.457130] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 BaseBdev1_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 [2024-09-30 14:11:56.027050] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:51.512 [2024-09-30 14:11:56.027205] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.512 [2024-09-30 14:11:56.027256] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:51.512 [2024-09-30 14:11:56.027296] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.512 [2024-09-30 14:11:56.029370] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.512 [2024-09-30 14:11:56.029442] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:51.512 BaseBdev1 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 BaseBdev2_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 [2024-09-30 14:11:56.070566] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:51.512 [2024-09-30 14:11:56.070703] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.512 [2024-09-30 14:11:56.070755] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:51.512 [2024-09-30 14:11:56.070786] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.512 [2024-09-30 14:11:56.075434] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.512 [2024-09-30 14:11:56.075518] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:51.512 BaseBdev2 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 BaseBdev3_malloc 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.512 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.512 [2024-09-30 14:11:56.101853] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:51.512 [2024-09-30 14:11:56.101974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.512 [2024-09-30 14:11:56.102015] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:51.513 [2024-09-30 14:11:56.102048] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.513 [2024-09-30 14:11:56.104112] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.513 [2024-09-30 14:11:56.104190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:51.513 BaseBdev3 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.513 BaseBdev4_malloc 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.513 [2024-09-30 14:11:56.130602] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:51.513 [2024-09-30 14:11:56.130704] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.513 [2024-09-30 14:11:56.130726] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:51.513 [2024-09-30 14:11:56.130736] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.513 [2024-09-30 14:11:56.132693] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.513 [2024-09-30 14:11:56.132729] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:51.513 BaseBdev4 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.513 spare_malloc 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.513 spare_delay 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.513 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.773 [2024-09-30 14:11:56.170961] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:51.773 [2024-09-30 14:11:56.171072] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.773 [2024-09-30 14:11:56.171119] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:51.773 [2024-09-30 14:11:56.171130] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.773 [2024-09-30 14:11:56.173142] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.773 [2024-09-30 14:11:56.173189] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:51.773 spare 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.773 [2024-09-30 14:11:56.183017] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:51.773 [2024-09-30 14:11:56.184871] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:51.773 [2024-09-30 14:11:56.184962] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:51.773 [2024-09-30 14:11:56.185026] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:51.773 [2024-09-30 14:11:56.185204] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:51.773 [2024-09-30 14:11:56.185250] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:51.773 [2024-09-30 14:11:56.185510] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:51.773 [2024-09-30 14:11:56.185678] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:51.773 [2024-09-30 14:11:56.185716] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:51.773 [2024-09-30 14:11:56.185858] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.773 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:51.773 "name": "raid_bdev1", 00:12:51.773 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:51.773 "strip_size_kb": 0, 00:12:51.773 "state": "online", 00:12:51.773 "raid_level": "raid1", 00:12:51.773 "superblock": true, 00:12:51.773 "num_base_bdevs": 4, 00:12:51.773 "num_base_bdevs_discovered": 4, 00:12:51.773 "num_base_bdevs_operational": 4, 00:12:51.773 "base_bdevs_list": [ 00:12:51.773 { 00:12:51.773 "name": "BaseBdev1", 00:12:51.773 "uuid": "d4ac6e53-9e1d-51ca-93bb-d821d7f38c70", 00:12:51.773 "is_configured": true, 00:12:51.773 "data_offset": 2048, 00:12:51.773 "data_size": 63488 00:12:51.773 }, 00:12:51.773 { 00:12:51.773 "name": "BaseBdev2", 00:12:51.773 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:51.773 "is_configured": true, 00:12:51.773 "data_offset": 2048, 00:12:51.773 "data_size": 63488 00:12:51.773 }, 00:12:51.773 { 00:12:51.773 "name": "BaseBdev3", 00:12:51.773 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:51.773 "is_configured": true, 00:12:51.773 "data_offset": 2048, 00:12:51.774 "data_size": 63488 00:12:51.774 }, 00:12:51.774 { 00:12:51.774 "name": "BaseBdev4", 00:12:51.774 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:51.774 "is_configured": true, 00:12:51.774 "data_offset": 2048, 00:12:51.774 "data_size": 63488 00:12:51.774 } 00:12:51.774 ] 00:12:51.774 }' 00:12:51.774 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:51.774 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.033 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:52.033 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:52.033 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.033 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.033 [2024-09-30 14:11:56.654462] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:52.034 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.294 [2024-09-30 14:11:56.737988] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.294 "name": "raid_bdev1", 00:12:52.294 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:52.294 "strip_size_kb": 0, 00:12:52.294 "state": "online", 00:12:52.294 "raid_level": "raid1", 00:12:52.294 "superblock": true, 00:12:52.294 "num_base_bdevs": 4, 00:12:52.294 "num_base_bdevs_discovered": 3, 00:12:52.294 "num_base_bdevs_operational": 3, 00:12:52.294 "base_bdevs_list": [ 00:12:52.294 { 00:12:52.294 "name": null, 00:12:52.294 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.294 "is_configured": false, 00:12:52.294 "data_offset": 0, 00:12:52.294 "data_size": 63488 00:12:52.294 }, 00:12:52.294 { 00:12:52.294 "name": "BaseBdev2", 00:12:52.294 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:52.294 "is_configured": true, 00:12:52.294 "data_offset": 2048, 00:12:52.294 "data_size": 63488 00:12:52.294 }, 00:12:52.294 { 00:12:52.294 "name": "BaseBdev3", 00:12:52.294 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:52.294 "is_configured": true, 00:12:52.294 "data_offset": 2048, 00:12:52.294 "data_size": 63488 00:12:52.294 }, 00:12:52.294 { 00:12:52.294 "name": "BaseBdev4", 00:12:52.294 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:52.294 "is_configured": true, 00:12:52.294 "data_offset": 2048, 00:12:52.294 "data_size": 63488 00:12:52.294 } 00:12:52.294 ] 00:12:52.294 }' 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.294 14:11:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.294 [2024-09-30 14:11:56.831732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:12:52.294 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:52.294 Zero copy mechanism will not be used. 00:12:52.294 Running I/O for 60 seconds... 00:12:52.554 14:11:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:52.554 14:11:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.554 14:11:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.554 [2024-09-30 14:11:57.168671] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:52.554 14:11:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.554 14:11:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:52.554 [2024-09-30 14:11:57.184122] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:12:52.554 [2024-09-30 14:11:57.186079] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:52.813 [2024-09-30 14:11:57.292946] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:52.813 [2024-09-30 14:11:57.294227] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:53.073 [2024-09-30 14:11:57.517475] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:53.073 [2024-09-30 14:11:57.518126] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:53.592 143.00 IOPS, 429.00 MiB/s [2024-09-30 14:11:57.989904] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:53.592 [2024-09-30 14:11:57.990603] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.592 "name": "raid_bdev1", 00:12:53.592 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:53.592 "strip_size_kb": 0, 00:12:53.592 "state": "online", 00:12:53.592 "raid_level": "raid1", 00:12:53.592 "superblock": true, 00:12:53.592 "num_base_bdevs": 4, 00:12:53.592 "num_base_bdevs_discovered": 4, 00:12:53.592 "num_base_bdevs_operational": 4, 00:12:53.592 "process": { 00:12:53.592 "type": "rebuild", 00:12:53.592 "target": "spare", 00:12:53.592 "progress": { 00:12:53.592 "blocks": 10240, 00:12:53.592 "percent": 16 00:12:53.592 } 00:12:53.592 }, 00:12:53.592 "base_bdevs_list": [ 00:12:53.592 { 00:12:53.592 "name": "spare", 00:12:53.592 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:53.592 "is_configured": true, 00:12:53.592 "data_offset": 2048, 00:12:53.592 "data_size": 63488 00:12:53.592 }, 00:12:53.592 { 00:12:53.592 "name": "BaseBdev2", 00:12:53.592 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:53.592 "is_configured": true, 00:12:53.592 "data_offset": 2048, 00:12:53.592 "data_size": 63488 00:12:53.592 }, 00:12:53.592 { 00:12:53.592 "name": "BaseBdev3", 00:12:53.592 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:53.592 "is_configured": true, 00:12:53.592 "data_offset": 2048, 00:12:53.592 "data_size": 63488 00:12:53.592 }, 00:12:53.592 { 00:12:53.592 "name": "BaseBdev4", 00:12:53.592 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:53.592 "is_configured": true, 00:12:53.592 "data_offset": 2048, 00:12:53.592 "data_size": 63488 00:12:53.592 } 00:12:53.592 ] 00:12:53.592 }' 00:12:53.592 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.852 [2024-09-30 14:11:58.331800] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.852 [2024-09-30 14:11:58.345064] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:53.852 [2024-09-30 14:11:58.352325] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:53.852 [2024-09-30 14:11:58.360121] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.852 [2024-09-30 14:11:58.360217] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.852 [2024-09-30 14:11:58.360233] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:53.852 [2024-09-30 14:11:58.381157] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000062f0 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.852 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.853 "name": "raid_bdev1", 00:12:53.853 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:53.853 "strip_size_kb": 0, 00:12:53.853 "state": "online", 00:12:53.853 "raid_level": "raid1", 00:12:53.853 "superblock": true, 00:12:53.853 "num_base_bdevs": 4, 00:12:53.853 "num_base_bdevs_discovered": 3, 00:12:53.853 "num_base_bdevs_operational": 3, 00:12:53.853 "base_bdevs_list": [ 00:12:53.853 { 00:12:53.853 "name": null, 00:12:53.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.853 "is_configured": false, 00:12:53.853 "data_offset": 0, 00:12:53.853 "data_size": 63488 00:12:53.853 }, 00:12:53.853 { 00:12:53.853 "name": "BaseBdev2", 00:12:53.853 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:53.853 "is_configured": true, 00:12:53.853 "data_offset": 2048, 00:12:53.853 "data_size": 63488 00:12:53.853 }, 00:12:53.853 { 00:12:53.853 "name": "BaseBdev3", 00:12:53.853 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:53.853 "is_configured": true, 00:12:53.853 "data_offset": 2048, 00:12:53.853 "data_size": 63488 00:12:53.853 }, 00:12:53.853 { 00:12:53.853 "name": "BaseBdev4", 00:12:53.853 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:53.853 "is_configured": true, 00:12:53.853 "data_offset": 2048, 00:12:53.853 "data_size": 63488 00:12:53.853 } 00:12:53.853 ] 00:12:53.853 }' 00:12:53.853 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.853 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:54.421 156.00 IOPS, 468.00 MiB/s 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:54.421 "name": "raid_bdev1", 00:12:54.421 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:54.421 "strip_size_kb": 0, 00:12:54.421 "state": "online", 00:12:54.421 "raid_level": "raid1", 00:12:54.421 "superblock": true, 00:12:54.421 "num_base_bdevs": 4, 00:12:54.421 "num_base_bdevs_discovered": 3, 00:12:54.421 "num_base_bdevs_operational": 3, 00:12:54.421 "base_bdevs_list": [ 00:12:54.421 { 00:12:54.421 "name": null, 00:12:54.421 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:54.421 "is_configured": false, 00:12:54.421 "data_offset": 0, 00:12:54.421 "data_size": 63488 00:12:54.421 }, 00:12:54.421 { 00:12:54.421 "name": "BaseBdev2", 00:12:54.421 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:54.421 "is_configured": true, 00:12:54.421 "data_offset": 2048, 00:12:54.421 "data_size": 63488 00:12:54.421 }, 00:12:54.421 { 00:12:54.421 "name": "BaseBdev3", 00:12:54.421 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:54.421 "is_configured": true, 00:12:54.421 "data_offset": 2048, 00:12:54.421 "data_size": 63488 00:12:54.421 }, 00:12:54.421 { 00:12:54.421 "name": "BaseBdev4", 00:12:54.421 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:54.421 "is_configured": true, 00:12:54.421 "data_offset": 2048, 00:12:54.421 "data_size": 63488 00:12:54.421 } 00:12:54.421 ] 00:12:54.421 }' 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.421 [2024-09-30 14:11:58.900978] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:54.421 14:11:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:54.421 [2024-09-30 14:11:58.958454] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:12:54.421 [2024-09-30 14:11:58.960389] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:54.421 [2024-09-30 14:11:59.068612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:54.421 [2024-09-30 14:11:59.069819] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:54.680 [2024-09-30 14:11:59.279050] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:54.680 [2024-09-30 14:11:59.279310] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:55.248 [2024-09-30 14:11:59.664305] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:55.248 [2024-09-30 14:11:59.665028] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:55.507 151.00 IOPS, 453.00 MiB/s 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.507 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.507 "name": "raid_bdev1", 00:12:55.507 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:55.507 "strip_size_kb": 0, 00:12:55.507 "state": "online", 00:12:55.507 "raid_level": "raid1", 00:12:55.507 "superblock": true, 00:12:55.507 "num_base_bdevs": 4, 00:12:55.507 "num_base_bdevs_discovered": 4, 00:12:55.507 "num_base_bdevs_operational": 4, 00:12:55.507 "process": { 00:12:55.507 "type": "rebuild", 00:12:55.507 "target": "spare", 00:12:55.507 "progress": { 00:12:55.507 "blocks": 12288, 00:12:55.507 "percent": 19 00:12:55.507 } 00:12:55.507 }, 00:12:55.507 "base_bdevs_list": [ 00:12:55.507 { 00:12:55.507 "name": "spare", 00:12:55.507 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:55.507 "is_configured": true, 00:12:55.507 "data_offset": 2048, 00:12:55.507 "data_size": 63488 00:12:55.507 }, 00:12:55.507 { 00:12:55.507 "name": "BaseBdev2", 00:12:55.507 "uuid": "34f1616b-7cb0-56dc-a669-b0afc3f93093", 00:12:55.508 "is_configured": true, 00:12:55.508 "data_offset": 2048, 00:12:55.508 "data_size": 63488 00:12:55.508 }, 00:12:55.508 { 00:12:55.508 "name": "BaseBdev3", 00:12:55.508 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:55.508 "is_configured": true, 00:12:55.508 "data_offset": 2048, 00:12:55.508 "data_size": 63488 00:12:55.508 }, 00:12:55.508 { 00:12:55.508 "name": "BaseBdev4", 00:12:55.508 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:55.508 "is_configured": true, 00:12:55.508 "data_offset": 2048, 00:12:55.508 "data_size": 63488 00:12:55.508 } 00:12:55.508 ] 00:12:55.508 }' 00:12:55.508 14:11:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:55.508 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.508 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.508 [2024-09-30 14:12:00.080224] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:55.508 [2024-09-30 14:12:00.144318] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:55.767 [2024-09-30 14:12:00.285482] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000062f0 00:12:55.767 [2024-09-30 14:12:00.285555] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006490 00:12:55.767 [2024-09-30 14:12:00.293174] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.768 "name": "raid_bdev1", 00:12:55.768 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:55.768 "strip_size_kb": 0, 00:12:55.768 "state": "online", 00:12:55.768 "raid_level": "raid1", 00:12:55.768 "superblock": true, 00:12:55.768 "num_base_bdevs": 4, 00:12:55.768 "num_base_bdevs_discovered": 3, 00:12:55.768 "num_base_bdevs_operational": 3, 00:12:55.768 "process": { 00:12:55.768 "type": "rebuild", 00:12:55.768 "target": "spare", 00:12:55.768 "progress": { 00:12:55.768 "blocks": 16384, 00:12:55.768 "percent": 25 00:12:55.768 } 00:12:55.768 }, 00:12:55.768 "base_bdevs_list": [ 00:12:55.768 { 00:12:55.768 "name": "spare", 00:12:55.768 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:55.768 "is_configured": true, 00:12:55.768 "data_offset": 2048, 00:12:55.768 "data_size": 63488 00:12:55.768 }, 00:12:55.768 { 00:12:55.768 "name": null, 00:12:55.768 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.768 "is_configured": false, 00:12:55.768 "data_offset": 0, 00:12:55.768 "data_size": 63488 00:12:55.768 }, 00:12:55.768 { 00:12:55.768 "name": "BaseBdev3", 00:12:55.768 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:55.768 "is_configured": true, 00:12:55.768 "data_offset": 2048, 00:12:55.768 "data_size": 63488 00:12:55.768 }, 00:12:55.768 { 00:12:55.768 "name": "BaseBdev4", 00:12:55.768 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:55.768 "is_configured": true, 00:12:55.768 "data_offset": 2048, 00:12:55.768 "data_size": 63488 00:12:55.768 } 00:12:55.768 ] 00:12:55.768 }' 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=401 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.768 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:56.028 "name": "raid_bdev1", 00:12:56.028 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:56.028 "strip_size_kb": 0, 00:12:56.028 "state": "online", 00:12:56.028 "raid_level": "raid1", 00:12:56.028 "superblock": true, 00:12:56.028 "num_base_bdevs": 4, 00:12:56.028 "num_base_bdevs_discovered": 3, 00:12:56.028 "num_base_bdevs_operational": 3, 00:12:56.028 "process": { 00:12:56.028 "type": "rebuild", 00:12:56.028 "target": "spare", 00:12:56.028 "progress": { 00:12:56.028 "blocks": 16384, 00:12:56.028 "percent": 25 00:12:56.028 } 00:12:56.028 }, 00:12:56.028 "base_bdevs_list": [ 00:12:56.028 { 00:12:56.028 "name": "spare", 00:12:56.028 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:56.028 "is_configured": true, 00:12:56.028 "data_offset": 2048, 00:12:56.028 "data_size": 63488 00:12:56.028 }, 00:12:56.028 { 00:12:56.028 "name": null, 00:12:56.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.028 "is_configured": false, 00:12:56.028 "data_offset": 0, 00:12:56.028 "data_size": 63488 00:12:56.028 }, 00:12:56.028 { 00:12:56.028 "name": "BaseBdev3", 00:12:56.028 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:56.028 "is_configured": true, 00:12:56.028 "data_offset": 2048, 00:12:56.028 "data_size": 63488 00:12:56.028 }, 00:12:56.028 { 00:12:56.028 "name": "BaseBdev4", 00:12:56.028 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:56.028 "is_configured": true, 00:12:56.028 "data_offset": 2048, 00:12:56.028 "data_size": 63488 00:12:56.028 } 00:12:56.028 ] 00:12:56.028 }' 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:56.028 14:12:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:56.856 129.25 IOPS, 387.75 MiB/s [2024-09-30 14:12:01.296336] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:56.856 [2024-09-30 14:12:01.402362] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:56.856 [2024-09-30 14:12:01.402787] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:57.116 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.117 "name": "raid_bdev1", 00:12:57.117 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:57.117 "strip_size_kb": 0, 00:12:57.117 "state": "online", 00:12:57.117 "raid_level": "raid1", 00:12:57.117 "superblock": true, 00:12:57.117 "num_base_bdevs": 4, 00:12:57.117 "num_base_bdevs_discovered": 3, 00:12:57.117 "num_base_bdevs_operational": 3, 00:12:57.117 "process": { 00:12:57.117 "type": "rebuild", 00:12:57.117 "target": "spare", 00:12:57.117 "progress": { 00:12:57.117 "blocks": 34816, 00:12:57.117 "percent": 54 00:12:57.117 } 00:12:57.117 }, 00:12:57.117 "base_bdevs_list": [ 00:12:57.117 { 00:12:57.117 "name": "spare", 00:12:57.117 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:57.117 "is_configured": true, 00:12:57.117 "data_offset": 2048, 00:12:57.117 "data_size": 63488 00:12:57.117 }, 00:12:57.117 { 00:12:57.117 "name": null, 00:12:57.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.117 "is_configured": false, 00:12:57.117 "data_offset": 0, 00:12:57.117 "data_size": 63488 00:12:57.117 }, 00:12:57.117 { 00:12:57.117 "name": "BaseBdev3", 00:12:57.117 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:57.117 "is_configured": true, 00:12:57.117 "data_offset": 2048, 00:12:57.117 "data_size": 63488 00:12:57.117 }, 00:12:57.117 { 00:12:57.117 "name": "BaseBdev4", 00:12:57.117 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:57.117 "is_configured": true, 00:12:57.117 "data_offset": 2048, 00:12:57.117 "data_size": 63488 00:12:57.117 } 00:12:57.117 ] 00:12:57.117 }' 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:57.117 14:12:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:57.376 112.00 IOPS, 336.00 MiB/s [2024-09-30 14:12:01.840944] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:57.636 [2024-09-30 14:12:02.163519] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:12:57.636 [2024-09-30 14:12:02.275449] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:58.205 [2024-09-30 14:12:02.586561] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:12:58.205 [2024-09-30 14:12:02.687533] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:58.205 "name": "raid_bdev1", 00:12:58.205 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:58.205 "strip_size_kb": 0, 00:12:58.205 "state": "online", 00:12:58.205 "raid_level": "raid1", 00:12:58.205 "superblock": true, 00:12:58.205 "num_base_bdevs": 4, 00:12:58.205 "num_base_bdevs_discovered": 3, 00:12:58.205 "num_base_bdevs_operational": 3, 00:12:58.205 "process": { 00:12:58.205 "type": "rebuild", 00:12:58.205 "target": "spare", 00:12:58.205 "progress": { 00:12:58.205 "blocks": 53248, 00:12:58.205 "percent": 83 00:12:58.205 } 00:12:58.205 }, 00:12:58.205 "base_bdevs_list": [ 00:12:58.205 { 00:12:58.205 "name": "spare", 00:12:58.205 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:58.205 "is_configured": true, 00:12:58.205 "data_offset": 2048, 00:12:58.205 "data_size": 63488 00:12:58.205 }, 00:12:58.205 { 00:12:58.205 "name": null, 00:12:58.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.205 "is_configured": false, 00:12:58.205 "data_offset": 0, 00:12:58.205 "data_size": 63488 00:12:58.205 }, 00:12:58.205 { 00:12:58.205 "name": "BaseBdev3", 00:12:58.205 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:58.205 "is_configured": true, 00:12:58.205 "data_offset": 2048, 00:12:58.205 "data_size": 63488 00:12:58.205 }, 00:12:58.205 { 00:12:58.205 "name": "BaseBdev4", 00:12:58.205 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:58.205 "is_configured": true, 00:12:58.205 "data_offset": 2048, 00:12:58.205 "data_size": 63488 00:12:58.205 } 00:12:58.205 ] 00:12:58.205 }' 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:58.205 14:12:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:58.465 100.83 IOPS, 302.50 MiB/s [2024-09-30 14:12:02.908675] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:12:58.723 [2024-09-30 14:12:03.236227] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:58.723 [2024-09-30 14:12:03.340952] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:58.723 [2024-09-30 14:12:03.344038] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.292 91.57 IOPS, 274.71 MiB/s 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.292 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.292 "name": "raid_bdev1", 00:12:59.292 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:59.292 "strip_size_kb": 0, 00:12:59.292 "state": "online", 00:12:59.292 "raid_level": "raid1", 00:12:59.292 "superblock": true, 00:12:59.292 "num_base_bdevs": 4, 00:12:59.292 "num_base_bdevs_discovered": 3, 00:12:59.292 "num_base_bdevs_operational": 3, 00:12:59.292 "base_bdevs_list": [ 00:12:59.292 { 00:12:59.292 "name": "spare", 00:12:59.292 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:59.292 "is_configured": true, 00:12:59.292 "data_offset": 2048, 00:12:59.292 "data_size": 63488 00:12:59.292 }, 00:12:59.292 { 00:12:59.292 "name": null, 00:12:59.292 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.292 "is_configured": false, 00:12:59.292 "data_offset": 0, 00:12:59.292 "data_size": 63488 00:12:59.292 }, 00:12:59.292 { 00:12:59.292 "name": "BaseBdev3", 00:12:59.292 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:59.292 "is_configured": true, 00:12:59.293 "data_offset": 2048, 00:12:59.293 "data_size": 63488 00:12:59.293 }, 00:12:59.293 { 00:12:59.293 "name": "BaseBdev4", 00:12:59.293 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:59.293 "is_configured": true, 00:12:59.293 "data_offset": 2048, 00:12:59.293 "data_size": 63488 00:12:59.293 } 00:12:59.293 ] 00:12:59.293 }' 00:12:59.293 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.293 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:59.293 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.553 "name": "raid_bdev1", 00:12:59.553 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:59.553 "strip_size_kb": 0, 00:12:59.553 "state": "online", 00:12:59.553 "raid_level": "raid1", 00:12:59.553 "superblock": true, 00:12:59.553 "num_base_bdevs": 4, 00:12:59.553 "num_base_bdevs_discovered": 3, 00:12:59.553 "num_base_bdevs_operational": 3, 00:12:59.553 "base_bdevs_list": [ 00:12:59.553 { 00:12:59.553 "name": "spare", 00:12:59.553 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": null, 00:12:59.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.553 "is_configured": false, 00:12:59.553 "data_offset": 0, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": "BaseBdev3", 00:12:59.553 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": "BaseBdev4", 00:12:59.553 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.553 } 00:12:59.553 ] 00:12:59.553 }' 00:12:59.553 14:12:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.553 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.553 "name": "raid_bdev1", 00:12:59.553 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:12:59.553 "strip_size_kb": 0, 00:12:59.553 "state": "online", 00:12:59.553 "raid_level": "raid1", 00:12:59.553 "superblock": true, 00:12:59.553 "num_base_bdevs": 4, 00:12:59.553 "num_base_bdevs_discovered": 3, 00:12:59.553 "num_base_bdevs_operational": 3, 00:12:59.553 "base_bdevs_list": [ 00:12:59.553 { 00:12:59.553 "name": "spare", 00:12:59.553 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": null, 00:12:59.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.553 "is_configured": false, 00:12:59.553 "data_offset": 0, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": "BaseBdev3", 00:12:59.553 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.553 }, 00:12:59.553 { 00:12:59.553 "name": "BaseBdev4", 00:12:59.553 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:12:59.553 "is_configured": true, 00:12:59.553 "data_offset": 2048, 00:12:59.553 "data_size": 63488 00:12:59.554 } 00:12:59.554 ] 00:12:59.554 }' 00:12:59.554 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.554 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.123 [2024-09-30 14:12:04.520598] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:00.123 [2024-09-30 14:12:04.520711] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:00.123 00:13:00.123 Latency(us) 00:13:00.123 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:00.123 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:00.123 raid_bdev1 : 7.71 86.90 260.71 0.00 0.00 15960.33 291.55 115389.15 00:13:00.123 =================================================================================================================== 00:13:00.123 Total : 86.90 260.71 0.00 0.00 15960.33 291.55 115389.15 00:13:00.123 [2024-09-30 14:12:04.531608] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.123 [2024-09-30 14:12:04.531680] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:00.123 [2024-09-30 14:12:04.531785] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:00.123 [2024-09-30 14:12:04.531853] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:00.123 { 00:13:00.123 "results": [ 00:13:00.123 { 00:13:00.123 "job": "raid_bdev1", 00:13:00.123 "core_mask": "0x1", 00:13:00.123 "workload": "randrw", 00:13:00.123 "percentage": 50, 00:13:00.123 "status": "finished", 00:13:00.123 "queue_depth": 2, 00:13:00.123 "io_size": 3145728, 00:13:00.123 "runtime": 7.709695, 00:13:00.123 "iops": 86.90356752115356, 00:13:00.123 "mibps": 260.7107025634607, 00:13:00.123 "io_failed": 0, 00:13:00.123 "io_timeout": 0, 00:13:00.123 "avg_latency_us": 15960.325756370981, 00:13:00.123 "min_latency_us": 291.54934497816595, 00:13:00.123 "max_latency_us": 115389.14934497817 00:13:00.123 } 00:13:00.123 ], 00:13:00.123 "core_count": 1 00:13:00.123 } 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.123 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:00.382 /dev/nbd0 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:00.383 1+0 records in 00:13:00.383 1+0 records out 00:13:00.383 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000500907 s, 8.2 MB/s 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.383 14:12:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:00.383 /dev/nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:00.643 1+0 records in 00:13:00.643 1+0 records out 00:13:00.643 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000634217 s, 6.5 MB/s 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:00.643 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.903 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:01.162 /dev/nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:01.163 1+0 records in 00:13:01.163 1+0 records out 00:13:01.163 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000376002 s, 10.9 MB/s 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:01.163 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:01.422 14:12:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.422 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.681 [2024-09-30 14:12:06.094534] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:01.681 [2024-09-30 14:12:06.094651] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:01.681 [2024-09-30 14:12:06.094688] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:13:01.681 [2024-09-30 14:12:06.094721] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:01.681 [2024-09-30 14:12:06.096950] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:01.681 [2024-09-30 14:12:06.097027] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:01.681 [2024-09-30 14:12:06.097129] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:01.681 [2024-09-30 14:12:06.097199] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:01.681 [2024-09-30 14:12:06.097328] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:01.681 [2024-09-30 14:12:06.097458] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:01.681 spare 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.681 [2024-09-30 14:12:06.197380] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:13:01.681 [2024-09-30 14:12:06.197461] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:01.681 [2024-09-30 14:12:06.197718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037230 00:13:01.681 [2024-09-30 14:12:06.197901] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:13:01.681 [2024-09-30 14:12:06.197944] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:13:01.681 [2024-09-30 14:12:06.198093] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.681 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:01.681 "name": "raid_bdev1", 00:13:01.681 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:01.681 "strip_size_kb": 0, 00:13:01.681 "state": "online", 00:13:01.681 "raid_level": "raid1", 00:13:01.681 "superblock": true, 00:13:01.681 "num_base_bdevs": 4, 00:13:01.682 "num_base_bdevs_discovered": 3, 00:13:01.682 "num_base_bdevs_operational": 3, 00:13:01.682 "base_bdevs_list": [ 00:13:01.682 { 00:13:01.682 "name": "spare", 00:13:01.682 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:13:01.682 "is_configured": true, 00:13:01.682 "data_offset": 2048, 00:13:01.682 "data_size": 63488 00:13:01.682 }, 00:13:01.682 { 00:13:01.682 "name": null, 00:13:01.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.682 "is_configured": false, 00:13:01.682 "data_offset": 2048, 00:13:01.682 "data_size": 63488 00:13:01.682 }, 00:13:01.682 { 00:13:01.682 "name": "BaseBdev3", 00:13:01.682 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:01.682 "is_configured": true, 00:13:01.682 "data_offset": 2048, 00:13:01.682 "data_size": 63488 00:13:01.682 }, 00:13:01.682 { 00:13:01.682 "name": "BaseBdev4", 00:13:01.682 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:01.682 "is_configured": true, 00:13:01.682 "data_offset": 2048, 00:13:01.682 "data_size": 63488 00:13:01.682 } 00:13:01.682 ] 00:13:01.682 }' 00:13:01.682 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:01.682 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.250 "name": "raid_bdev1", 00:13:02.250 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:02.250 "strip_size_kb": 0, 00:13:02.250 "state": "online", 00:13:02.250 "raid_level": "raid1", 00:13:02.250 "superblock": true, 00:13:02.250 "num_base_bdevs": 4, 00:13:02.250 "num_base_bdevs_discovered": 3, 00:13:02.250 "num_base_bdevs_operational": 3, 00:13:02.250 "base_bdevs_list": [ 00:13:02.250 { 00:13:02.250 "name": "spare", 00:13:02.250 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:13:02.250 "is_configured": true, 00:13:02.250 "data_offset": 2048, 00:13:02.250 "data_size": 63488 00:13:02.250 }, 00:13:02.250 { 00:13:02.250 "name": null, 00:13:02.250 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.250 "is_configured": false, 00:13:02.250 "data_offset": 2048, 00:13:02.250 "data_size": 63488 00:13:02.250 }, 00:13:02.250 { 00:13:02.250 "name": "BaseBdev3", 00:13:02.250 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:02.250 "is_configured": true, 00:13:02.250 "data_offset": 2048, 00:13:02.250 "data_size": 63488 00:13:02.250 }, 00:13:02.250 { 00:13:02.250 "name": "BaseBdev4", 00:13:02.250 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:02.250 "is_configured": true, 00:13:02.250 "data_offset": 2048, 00:13:02.250 "data_size": 63488 00:13:02.250 } 00:13:02.250 ] 00:13:02.250 }' 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.250 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.251 [2024-09-30 14:12:06.841362] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.251 "name": "raid_bdev1", 00:13:02.251 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:02.251 "strip_size_kb": 0, 00:13:02.251 "state": "online", 00:13:02.251 "raid_level": "raid1", 00:13:02.251 "superblock": true, 00:13:02.251 "num_base_bdevs": 4, 00:13:02.251 "num_base_bdevs_discovered": 2, 00:13:02.251 "num_base_bdevs_operational": 2, 00:13:02.251 "base_bdevs_list": [ 00:13:02.251 { 00:13:02.251 "name": null, 00:13:02.251 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.251 "is_configured": false, 00:13:02.251 "data_offset": 0, 00:13:02.251 "data_size": 63488 00:13:02.251 }, 00:13:02.251 { 00:13:02.251 "name": null, 00:13:02.251 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.251 "is_configured": false, 00:13:02.251 "data_offset": 2048, 00:13:02.251 "data_size": 63488 00:13:02.251 }, 00:13:02.251 { 00:13:02.251 "name": "BaseBdev3", 00:13:02.251 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:02.251 "is_configured": true, 00:13:02.251 "data_offset": 2048, 00:13:02.251 "data_size": 63488 00:13:02.251 }, 00:13:02.251 { 00:13:02.251 "name": "BaseBdev4", 00:13:02.251 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:02.251 "is_configured": true, 00:13:02.251 "data_offset": 2048, 00:13:02.251 "data_size": 63488 00:13:02.251 } 00:13:02.251 ] 00:13:02.251 }' 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.251 14:12:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.819 14:12:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:02.819 14:12:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.819 14:12:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.819 [2024-09-30 14:12:07.304603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:02.819 [2024-09-30 14:12:07.304783] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:02.819 [2024-09-30 14:12:07.304847] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:02.819 [2024-09-30 14:12:07.304934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:02.819 [2024-09-30 14:12:07.308427] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037300 00:13:02.819 [2024-09-30 14:12:07.310198] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:02.819 14:12:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.819 14:12:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.756 "name": "raid_bdev1", 00:13:03.756 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:03.756 "strip_size_kb": 0, 00:13:03.756 "state": "online", 00:13:03.756 "raid_level": "raid1", 00:13:03.756 "superblock": true, 00:13:03.756 "num_base_bdevs": 4, 00:13:03.756 "num_base_bdevs_discovered": 3, 00:13:03.756 "num_base_bdevs_operational": 3, 00:13:03.756 "process": { 00:13:03.756 "type": "rebuild", 00:13:03.756 "target": "spare", 00:13:03.756 "progress": { 00:13:03.756 "blocks": 20480, 00:13:03.756 "percent": 32 00:13:03.756 } 00:13:03.756 }, 00:13:03.756 "base_bdevs_list": [ 00:13:03.756 { 00:13:03.756 "name": "spare", 00:13:03.756 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:13:03.756 "is_configured": true, 00:13:03.756 "data_offset": 2048, 00:13:03.756 "data_size": 63488 00:13:03.756 }, 00:13:03.756 { 00:13:03.756 "name": null, 00:13:03.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.756 "is_configured": false, 00:13:03.756 "data_offset": 2048, 00:13:03.756 "data_size": 63488 00:13:03.756 }, 00:13:03.756 { 00:13:03.756 "name": "BaseBdev3", 00:13:03.756 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:03.756 "is_configured": true, 00:13:03.756 "data_offset": 2048, 00:13:03.756 "data_size": 63488 00:13:03.756 }, 00:13:03.756 { 00:13:03.756 "name": "BaseBdev4", 00:13:03.756 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:03.756 "is_configured": true, 00:13:03.756 "data_offset": 2048, 00:13:03.756 "data_size": 63488 00:13:03.756 } 00:13:03.756 ] 00:13:03.756 }' 00:13:03.756 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.015 [2024-09-30 14:12:08.472929] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:04.015 [2024-09-30 14:12:08.514100] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:04.015 [2024-09-30 14:12:08.514208] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:04.015 [2024-09-30 14:12:08.514260] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:04.015 [2024-09-30 14:12:08.514272] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:04.015 "name": "raid_bdev1", 00:13:04.015 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:04.015 "strip_size_kb": 0, 00:13:04.015 "state": "online", 00:13:04.015 "raid_level": "raid1", 00:13:04.015 "superblock": true, 00:13:04.015 "num_base_bdevs": 4, 00:13:04.015 "num_base_bdevs_discovered": 2, 00:13:04.015 "num_base_bdevs_operational": 2, 00:13:04.015 "base_bdevs_list": [ 00:13:04.015 { 00:13:04.015 "name": null, 00:13:04.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.015 "is_configured": false, 00:13:04.015 "data_offset": 0, 00:13:04.015 "data_size": 63488 00:13:04.015 }, 00:13:04.015 { 00:13:04.015 "name": null, 00:13:04.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.015 "is_configured": false, 00:13:04.015 "data_offset": 2048, 00:13:04.015 "data_size": 63488 00:13:04.015 }, 00:13:04.015 { 00:13:04.015 "name": "BaseBdev3", 00:13:04.015 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:04.015 "is_configured": true, 00:13:04.015 "data_offset": 2048, 00:13:04.015 "data_size": 63488 00:13:04.015 }, 00:13:04.015 { 00:13:04.015 "name": "BaseBdev4", 00:13:04.015 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:04.015 "is_configured": true, 00:13:04.015 "data_offset": 2048, 00:13:04.015 "data_size": 63488 00:13:04.015 } 00:13:04.015 ] 00:13:04.015 }' 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:04.015 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.582 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:04.582 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.582 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.582 [2024-09-30 14:12:08.989374] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:04.582 [2024-09-30 14:12:08.989475] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:04.582 [2024-09-30 14:12:08.989517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:13:04.582 [2024-09-30 14:12:08.989546] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:04.583 [2024-09-30 14:12:08.989963] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:04.583 [2024-09-30 14:12:08.990023] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:04.583 [2024-09-30 14:12:08.990121] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:04.583 [2024-09-30 14:12:08.990163] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:04.583 [2024-09-30 14:12:08.990207] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:04.583 [2024-09-30 14:12:08.990259] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:04.583 [2024-09-30 14:12:08.993367] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000373d0 00:13:04.583 spare 00:13:04.583 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.583 [2024-09-30 14:12:08.995099] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:04.583 14:12:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:05.519 14:12:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.519 14:12:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.519 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.519 "name": "raid_bdev1", 00:13:05.519 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:05.519 "strip_size_kb": 0, 00:13:05.519 "state": "online", 00:13:05.519 "raid_level": "raid1", 00:13:05.519 "superblock": true, 00:13:05.519 "num_base_bdevs": 4, 00:13:05.519 "num_base_bdevs_discovered": 3, 00:13:05.519 "num_base_bdevs_operational": 3, 00:13:05.519 "process": { 00:13:05.519 "type": "rebuild", 00:13:05.519 "target": "spare", 00:13:05.519 "progress": { 00:13:05.519 "blocks": 20480, 00:13:05.519 "percent": 32 00:13:05.519 } 00:13:05.519 }, 00:13:05.519 "base_bdevs_list": [ 00:13:05.519 { 00:13:05.519 "name": "spare", 00:13:05.519 "uuid": "81fa9184-7090-52ec-9725-6c3b43946ce7", 00:13:05.519 "is_configured": true, 00:13:05.519 "data_offset": 2048, 00:13:05.519 "data_size": 63488 00:13:05.519 }, 00:13:05.519 { 00:13:05.519 "name": null, 00:13:05.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.519 "is_configured": false, 00:13:05.519 "data_offset": 2048, 00:13:05.519 "data_size": 63488 00:13:05.519 }, 00:13:05.519 { 00:13:05.519 "name": "BaseBdev3", 00:13:05.519 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:05.519 "is_configured": true, 00:13:05.519 "data_offset": 2048, 00:13:05.519 "data_size": 63488 00:13:05.519 }, 00:13:05.519 { 00:13:05.519 "name": "BaseBdev4", 00:13:05.519 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:05.520 "is_configured": true, 00:13:05.520 "data_offset": 2048, 00:13:05.520 "data_size": 63488 00:13:05.520 } 00:13:05.520 ] 00:13:05.520 }' 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.520 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.520 [2024-09-30 14:12:10.156035] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:05.779 [2024-09-30 14:12:10.198930] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:05.779 [2024-09-30 14:12:10.199030] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:05.779 [2024-09-30 14:12:10.199063] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:05.779 [2024-09-30 14:12:10.199073] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:05.779 "name": "raid_bdev1", 00:13:05.779 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:05.779 "strip_size_kb": 0, 00:13:05.779 "state": "online", 00:13:05.779 "raid_level": "raid1", 00:13:05.779 "superblock": true, 00:13:05.779 "num_base_bdevs": 4, 00:13:05.779 "num_base_bdevs_discovered": 2, 00:13:05.779 "num_base_bdevs_operational": 2, 00:13:05.779 "base_bdevs_list": [ 00:13:05.779 { 00:13:05.779 "name": null, 00:13:05.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.779 "is_configured": false, 00:13:05.779 "data_offset": 0, 00:13:05.779 "data_size": 63488 00:13:05.779 }, 00:13:05.779 { 00:13:05.779 "name": null, 00:13:05.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.779 "is_configured": false, 00:13:05.779 "data_offset": 2048, 00:13:05.779 "data_size": 63488 00:13:05.779 }, 00:13:05.779 { 00:13:05.779 "name": "BaseBdev3", 00:13:05.779 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:05.779 "is_configured": true, 00:13:05.779 "data_offset": 2048, 00:13:05.779 "data_size": 63488 00:13:05.779 }, 00:13:05.779 { 00:13:05.779 "name": "BaseBdev4", 00:13:05.779 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:05.779 "is_configured": true, 00:13:05.779 "data_offset": 2048, 00:13:05.779 "data_size": 63488 00:13:05.779 } 00:13:05.779 ] 00:13:05.779 }' 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:05.779 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.039 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.298 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.298 "name": "raid_bdev1", 00:13:06.298 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:06.298 "strip_size_kb": 0, 00:13:06.298 "state": "online", 00:13:06.298 "raid_level": "raid1", 00:13:06.298 "superblock": true, 00:13:06.298 "num_base_bdevs": 4, 00:13:06.298 "num_base_bdevs_discovered": 2, 00:13:06.298 "num_base_bdevs_operational": 2, 00:13:06.298 "base_bdevs_list": [ 00:13:06.298 { 00:13:06.298 "name": null, 00:13:06.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.298 "is_configured": false, 00:13:06.298 "data_offset": 0, 00:13:06.298 "data_size": 63488 00:13:06.298 }, 00:13:06.298 { 00:13:06.298 "name": null, 00:13:06.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.298 "is_configured": false, 00:13:06.298 "data_offset": 2048, 00:13:06.298 "data_size": 63488 00:13:06.298 }, 00:13:06.298 { 00:13:06.298 "name": "BaseBdev3", 00:13:06.298 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:06.298 "is_configured": true, 00:13:06.298 "data_offset": 2048, 00:13:06.298 "data_size": 63488 00:13:06.298 }, 00:13:06.298 { 00:13:06.298 "name": "BaseBdev4", 00:13:06.298 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:06.298 "is_configured": true, 00:13:06.298 "data_offset": 2048, 00:13:06.298 "data_size": 63488 00:13:06.298 } 00:13:06.299 ] 00:13:06.299 }' 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.299 [2024-09-30 14:12:10.793851] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:06.299 [2024-09-30 14:12:10.793948] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:06.299 [2024-09-30 14:12:10.793985] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:13:06.299 [2024-09-30 14:12:10.794012] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:06.299 [2024-09-30 14:12:10.794375] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:06.299 [2024-09-30 14:12:10.794393] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:06.299 [2024-09-30 14:12:10.794458] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:06.299 [2024-09-30 14:12:10.794472] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:06.299 [2024-09-30 14:12:10.794492] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:06.299 [2024-09-30 14:12:10.794502] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:06.299 BaseBdev1 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.299 14:12:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.236 "name": "raid_bdev1", 00:13:07.236 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:07.236 "strip_size_kb": 0, 00:13:07.236 "state": "online", 00:13:07.236 "raid_level": "raid1", 00:13:07.236 "superblock": true, 00:13:07.236 "num_base_bdevs": 4, 00:13:07.236 "num_base_bdevs_discovered": 2, 00:13:07.236 "num_base_bdevs_operational": 2, 00:13:07.236 "base_bdevs_list": [ 00:13:07.236 { 00:13:07.236 "name": null, 00:13:07.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.236 "is_configured": false, 00:13:07.236 "data_offset": 0, 00:13:07.236 "data_size": 63488 00:13:07.236 }, 00:13:07.236 { 00:13:07.236 "name": null, 00:13:07.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.236 "is_configured": false, 00:13:07.236 "data_offset": 2048, 00:13:07.236 "data_size": 63488 00:13:07.236 }, 00:13:07.236 { 00:13:07.236 "name": "BaseBdev3", 00:13:07.236 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:07.236 "is_configured": true, 00:13:07.236 "data_offset": 2048, 00:13:07.236 "data_size": 63488 00:13:07.236 }, 00:13:07.236 { 00:13:07.236 "name": "BaseBdev4", 00:13:07.236 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:07.236 "is_configured": true, 00:13:07.236 "data_offset": 2048, 00:13:07.236 "data_size": 63488 00:13:07.236 } 00:13:07.236 ] 00:13:07.236 }' 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.236 14:12:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.806 "name": "raid_bdev1", 00:13:07.806 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:07.806 "strip_size_kb": 0, 00:13:07.806 "state": "online", 00:13:07.806 "raid_level": "raid1", 00:13:07.806 "superblock": true, 00:13:07.806 "num_base_bdevs": 4, 00:13:07.806 "num_base_bdevs_discovered": 2, 00:13:07.806 "num_base_bdevs_operational": 2, 00:13:07.806 "base_bdevs_list": [ 00:13:07.806 { 00:13:07.806 "name": null, 00:13:07.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.806 "is_configured": false, 00:13:07.806 "data_offset": 0, 00:13:07.806 "data_size": 63488 00:13:07.806 }, 00:13:07.806 { 00:13:07.806 "name": null, 00:13:07.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.806 "is_configured": false, 00:13:07.806 "data_offset": 2048, 00:13:07.806 "data_size": 63488 00:13:07.806 }, 00:13:07.806 { 00:13:07.806 "name": "BaseBdev3", 00:13:07.806 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:07.806 "is_configured": true, 00:13:07.806 "data_offset": 2048, 00:13:07.806 "data_size": 63488 00:13:07.806 }, 00:13:07.806 { 00:13:07.806 "name": "BaseBdev4", 00:13:07.806 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:07.806 "is_configured": true, 00:13:07.806 "data_offset": 2048, 00:13:07.806 "data_size": 63488 00:13:07.806 } 00:13:07.806 ] 00:13:07.806 }' 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.806 [2024-09-30 14:12:12.363330] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:07.806 [2024-09-30 14:12:12.363519] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:07.806 [2024-09-30 14:12:12.363574] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:07.806 request: 00:13:07.806 { 00:13:07.806 "base_bdev": "BaseBdev1", 00:13:07.806 "raid_bdev": "raid_bdev1", 00:13:07.806 "method": "bdev_raid_add_base_bdev", 00:13:07.806 "req_id": 1 00:13:07.806 } 00:13:07.806 Got JSON-RPC error response 00:13:07.806 response: 00:13:07.806 { 00:13:07.806 "code": -22, 00:13:07.806 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:07.806 } 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:07.806 14:12:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.745 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.005 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.005 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.005 "name": "raid_bdev1", 00:13:09.005 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:09.005 "strip_size_kb": 0, 00:13:09.005 "state": "online", 00:13:09.005 "raid_level": "raid1", 00:13:09.005 "superblock": true, 00:13:09.005 "num_base_bdevs": 4, 00:13:09.005 "num_base_bdevs_discovered": 2, 00:13:09.005 "num_base_bdevs_operational": 2, 00:13:09.005 "base_bdevs_list": [ 00:13:09.005 { 00:13:09.005 "name": null, 00:13:09.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.005 "is_configured": false, 00:13:09.005 "data_offset": 0, 00:13:09.005 "data_size": 63488 00:13:09.005 }, 00:13:09.005 { 00:13:09.005 "name": null, 00:13:09.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.005 "is_configured": false, 00:13:09.005 "data_offset": 2048, 00:13:09.005 "data_size": 63488 00:13:09.005 }, 00:13:09.005 { 00:13:09.005 "name": "BaseBdev3", 00:13:09.005 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:09.005 "is_configured": true, 00:13:09.005 "data_offset": 2048, 00:13:09.005 "data_size": 63488 00:13:09.005 }, 00:13:09.005 { 00:13:09.005 "name": "BaseBdev4", 00:13:09.005 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:09.005 "is_configured": true, 00:13:09.005 "data_offset": 2048, 00:13:09.005 "data_size": 63488 00:13:09.005 } 00:13:09.005 ] 00:13:09.005 }' 00:13:09.005 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.005 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.265 "name": "raid_bdev1", 00:13:09.265 "uuid": "7db03c3e-871a-44fa-8bf2-4fb7923ac091", 00:13:09.265 "strip_size_kb": 0, 00:13:09.265 "state": "online", 00:13:09.265 "raid_level": "raid1", 00:13:09.265 "superblock": true, 00:13:09.265 "num_base_bdevs": 4, 00:13:09.265 "num_base_bdevs_discovered": 2, 00:13:09.265 "num_base_bdevs_operational": 2, 00:13:09.265 "base_bdevs_list": [ 00:13:09.265 { 00:13:09.265 "name": null, 00:13:09.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.265 "is_configured": false, 00:13:09.265 "data_offset": 0, 00:13:09.265 "data_size": 63488 00:13:09.265 }, 00:13:09.265 { 00:13:09.265 "name": null, 00:13:09.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.265 "is_configured": false, 00:13:09.265 "data_offset": 2048, 00:13:09.265 "data_size": 63488 00:13:09.265 }, 00:13:09.265 { 00:13:09.265 "name": "BaseBdev3", 00:13:09.265 "uuid": "189a3575-27f8-563a-8ceb-a76336f5664d", 00:13:09.265 "is_configured": true, 00:13:09.265 "data_offset": 2048, 00:13:09.265 "data_size": 63488 00:13:09.265 }, 00:13:09.265 { 00:13:09.265 "name": "BaseBdev4", 00:13:09.265 "uuid": "3ef9ced3-68e8-52be-bc17-654048f2d354", 00:13:09.265 "is_configured": true, 00:13:09.265 "data_offset": 2048, 00:13:09.265 "data_size": 63488 00:13:09.265 } 00:13:09.265 ] 00:13:09.265 }' 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:09.265 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 90768 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 90768 ']' 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 90768 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:09.526 14:12:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90768 00:13:09.526 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:09.526 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:09.526 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90768' 00:13:09.526 killing process with pid 90768 00:13:09.526 Received shutdown signal, test time was about 17.211334 seconds 00:13:09.526 00:13:09.526 Latency(us) 00:13:09.526 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:09.526 =================================================================================================================== 00:13:09.526 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:09.526 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 90768 00:13:09.526 [2024-09-30 14:12:14.011878] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:09.526 [2024-09-30 14:12:14.011994] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:09.526 [2024-09-30 14:12:14.012054] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:09.526 [2024-09-30 14:12:14.012065] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:13:09.526 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 90768 00:13:09.526 [2024-09-30 14:12:14.056014] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:09.786 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:09.786 00:13:09.786 real 0m19.194s 00:13:09.786 user 0m25.504s 00:13:09.786 sys 0m2.521s 00:13:09.786 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:09.786 ************************************ 00:13:09.786 END TEST raid_rebuild_test_sb_io 00:13:09.786 ************************************ 00:13:09.787 14:12:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.787 14:12:14 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:09.787 14:12:14 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:09.787 14:12:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:09.787 14:12:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:09.787 14:12:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:09.787 ************************************ 00:13:09.787 START TEST raid5f_state_function_test 00:13:09.787 ************************************ 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 false 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=91474 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 91474' 00:13:09.787 Process raid pid: 91474 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 91474 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 91474 ']' 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:09.787 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:09.787 14:12:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.047 [2024-09-30 14:12:14.474751] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:13:10.047 [2024-09-30 14:12:14.474975] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:10.047 [2024-09-30 14:12:14.608432] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:13:10.047 [2024-09-30 14:12:14.638446] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:10.047 [2024-09-30 14:12:14.684650] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:10.307 [2024-09-30 14:12:14.727055] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:10.307 [2024-09-30 14:12:14.727092] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:10.876 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.877 [2024-09-30 14:12:15.284271] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:10.877 [2024-09-30 14:12:15.284399] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:10.877 [2024-09-30 14:12:15.284432] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:10.877 [2024-09-30 14:12:15.284453] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:10.877 [2024-09-30 14:12:15.284476] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:10.877 [2024-09-30 14:12:15.284510] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.877 "name": "Existed_Raid", 00:13:10.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.877 "strip_size_kb": 64, 00:13:10.877 "state": "configuring", 00:13:10.877 "raid_level": "raid5f", 00:13:10.877 "superblock": false, 00:13:10.877 "num_base_bdevs": 3, 00:13:10.877 "num_base_bdevs_discovered": 0, 00:13:10.877 "num_base_bdevs_operational": 3, 00:13:10.877 "base_bdevs_list": [ 00:13:10.877 { 00:13:10.877 "name": "BaseBdev1", 00:13:10.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.877 "is_configured": false, 00:13:10.877 "data_offset": 0, 00:13:10.877 "data_size": 0 00:13:10.877 }, 00:13:10.877 { 00:13:10.877 "name": "BaseBdev2", 00:13:10.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.877 "is_configured": false, 00:13:10.877 "data_offset": 0, 00:13:10.877 "data_size": 0 00:13:10.877 }, 00:13:10.877 { 00:13:10.877 "name": "BaseBdev3", 00:13:10.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.877 "is_configured": false, 00:13:10.877 "data_offset": 0, 00:13:10.877 "data_size": 0 00:13:10.877 } 00:13:10.877 ] 00:13:10.877 }' 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.877 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 [2024-09-30 14:12:15.703477] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:11.137 [2024-09-30 14:12:15.703572] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 [2024-09-30 14:12:15.715467] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:11.137 [2024-09-30 14:12:15.715552] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:11.137 [2024-09-30 14:12:15.715579] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:11.137 [2024-09-30 14:12:15.715598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:11.137 [2024-09-30 14:12:15.715616] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:11.137 [2024-09-30 14:12:15.715633] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 [2024-09-30 14:12:15.736220] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:11.137 BaseBdev1 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.137 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.137 [ 00:13:11.137 { 00:13:11.137 "name": "BaseBdev1", 00:13:11.137 "aliases": [ 00:13:11.137 "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada" 00:13:11.137 ], 00:13:11.137 "product_name": "Malloc disk", 00:13:11.137 "block_size": 512, 00:13:11.137 "num_blocks": 65536, 00:13:11.137 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:11.137 "assigned_rate_limits": { 00:13:11.137 "rw_ios_per_sec": 0, 00:13:11.137 "rw_mbytes_per_sec": 0, 00:13:11.137 "r_mbytes_per_sec": 0, 00:13:11.137 "w_mbytes_per_sec": 0 00:13:11.137 }, 00:13:11.138 "claimed": true, 00:13:11.138 "claim_type": "exclusive_write", 00:13:11.138 "zoned": false, 00:13:11.138 "supported_io_types": { 00:13:11.138 "read": true, 00:13:11.138 "write": true, 00:13:11.138 "unmap": true, 00:13:11.138 "flush": true, 00:13:11.138 "reset": true, 00:13:11.138 "nvme_admin": false, 00:13:11.138 "nvme_io": false, 00:13:11.138 "nvme_io_md": false, 00:13:11.138 "write_zeroes": true, 00:13:11.138 "zcopy": true, 00:13:11.138 "get_zone_info": false, 00:13:11.138 "zone_management": false, 00:13:11.138 "zone_append": false, 00:13:11.138 "compare": false, 00:13:11.138 "compare_and_write": false, 00:13:11.138 "abort": true, 00:13:11.138 "seek_hole": false, 00:13:11.138 "seek_data": false, 00:13:11.138 "copy": true, 00:13:11.138 "nvme_iov_md": false 00:13:11.138 }, 00:13:11.138 "memory_domains": [ 00:13:11.138 { 00:13:11.138 "dma_device_id": "system", 00:13:11.138 "dma_device_type": 1 00:13:11.138 }, 00:13:11.138 { 00:13:11.138 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:11.138 "dma_device_type": 2 00:13:11.138 } 00:13:11.138 ], 00:13:11.138 "driver_specific": {} 00:13:11.138 } 00:13:11.138 ] 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.138 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.398 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.398 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.398 "name": "Existed_Raid", 00:13:11.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.398 "strip_size_kb": 64, 00:13:11.398 "state": "configuring", 00:13:11.398 "raid_level": "raid5f", 00:13:11.398 "superblock": false, 00:13:11.398 "num_base_bdevs": 3, 00:13:11.398 "num_base_bdevs_discovered": 1, 00:13:11.398 "num_base_bdevs_operational": 3, 00:13:11.398 "base_bdevs_list": [ 00:13:11.398 { 00:13:11.398 "name": "BaseBdev1", 00:13:11.398 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:11.398 "is_configured": true, 00:13:11.398 "data_offset": 0, 00:13:11.398 "data_size": 65536 00:13:11.398 }, 00:13:11.398 { 00:13:11.398 "name": "BaseBdev2", 00:13:11.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.398 "is_configured": false, 00:13:11.398 "data_offset": 0, 00:13:11.398 "data_size": 0 00:13:11.398 }, 00:13:11.398 { 00:13:11.398 "name": "BaseBdev3", 00:13:11.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.398 "is_configured": false, 00:13:11.398 "data_offset": 0, 00:13:11.398 "data_size": 0 00:13:11.398 } 00:13:11.398 ] 00:13:11.398 }' 00:13:11.398 14:12:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.398 14:12:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.658 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.659 [2024-09-30 14:12:16.239435] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:11.659 [2024-09-30 14:12:16.239557] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.659 [2024-09-30 14:12:16.251432] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:11.659 [2024-09-30 14:12:16.253204] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:11.659 [2024-09-30 14:12:16.253270] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:11.659 [2024-09-30 14:12:16.253299] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:11.659 [2024-09-30 14:12:16.253319] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.659 "name": "Existed_Raid", 00:13:11.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.659 "strip_size_kb": 64, 00:13:11.659 "state": "configuring", 00:13:11.659 "raid_level": "raid5f", 00:13:11.659 "superblock": false, 00:13:11.659 "num_base_bdevs": 3, 00:13:11.659 "num_base_bdevs_discovered": 1, 00:13:11.659 "num_base_bdevs_operational": 3, 00:13:11.659 "base_bdevs_list": [ 00:13:11.659 { 00:13:11.659 "name": "BaseBdev1", 00:13:11.659 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:11.659 "is_configured": true, 00:13:11.659 "data_offset": 0, 00:13:11.659 "data_size": 65536 00:13:11.659 }, 00:13:11.659 { 00:13:11.659 "name": "BaseBdev2", 00:13:11.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.659 "is_configured": false, 00:13:11.659 "data_offset": 0, 00:13:11.659 "data_size": 0 00:13:11.659 }, 00:13:11.659 { 00:13:11.659 "name": "BaseBdev3", 00:13:11.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.659 "is_configured": false, 00:13:11.659 "data_offset": 0, 00:13:11.659 "data_size": 0 00:13:11.659 } 00:13:11.659 ] 00:13:11.659 }' 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.659 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.229 [2024-09-30 14:12:16.735652] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:12.229 BaseBdev2 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.229 [ 00:13:12.229 { 00:13:12.229 "name": "BaseBdev2", 00:13:12.229 "aliases": [ 00:13:12.229 "d94879de-263f-438d-8291-28c995ce7f39" 00:13:12.229 ], 00:13:12.229 "product_name": "Malloc disk", 00:13:12.229 "block_size": 512, 00:13:12.229 "num_blocks": 65536, 00:13:12.229 "uuid": "d94879de-263f-438d-8291-28c995ce7f39", 00:13:12.229 "assigned_rate_limits": { 00:13:12.229 "rw_ios_per_sec": 0, 00:13:12.229 "rw_mbytes_per_sec": 0, 00:13:12.229 "r_mbytes_per_sec": 0, 00:13:12.229 "w_mbytes_per_sec": 0 00:13:12.229 }, 00:13:12.229 "claimed": true, 00:13:12.229 "claim_type": "exclusive_write", 00:13:12.229 "zoned": false, 00:13:12.229 "supported_io_types": { 00:13:12.229 "read": true, 00:13:12.229 "write": true, 00:13:12.229 "unmap": true, 00:13:12.229 "flush": true, 00:13:12.229 "reset": true, 00:13:12.229 "nvme_admin": false, 00:13:12.229 "nvme_io": false, 00:13:12.229 "nvme_io_md": false, 00:13:12.229 "write_zeroes": true, 00:13:12.229 "zcopy": true, 00:13:12.229 "get_zone_info": false, 00:13:12.229 "zone_management": false, 00:13:12.229 "zone_append": false, 00:13:12.229 "compare": false, 00:13:12.229 "compare_and_write": false, 00:13:12.229 "abort": true, 00:13:12.229 "seek_hole": false, 00:13:12.229 "seek_data": false, 00:13:12.229 "copy": true, 00:13:12.229 "nvme_iov_md": false 00:13:12.229 }, 00:13:12.229 "memory_domains": [ 00:13:12.229 { 00:13:12.229 "dma_device_id": "system", 00:13:12.229 "dma_device_type": 1 00:13:12.229 }, 00:13:12.229 { 00:13:12.229 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:12.229 "dma_device_type": 2 00:13:12.229 } 00:13:12.229 ], 00:13:12.229 "driver_specific": {} 00:13:12.229 } 00:13:12.229 ] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.229 "name": "Existed_Raid", 00:13:12.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.229 "strip_size_kb": 64, 00:13:12.229 "state": "configuring", 00:13:12.229 "raid_level": "raid5f", 00:13:12.229 "superblock": false, 00:13:12.229 "num_base_bdevs": 3, 00:13:12.229 "num_base_bdevs_discovered": 2, 00:13:12.229 "num_base_bdevs_operational": 3, 00:13:12.229 "base_bdevs_list": [ 00:13:12.229 { 00:13:12.229 "name": "BaseBdev1", 00:13:12.229 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:12.229 "is_configured": true, 00:13:12.229 "data_offset": 0, 00:13:12.229 "data_size": 65536 00:13:12.229 }, 00:13:12.229 { 00:13:12.229 "name": "BaseBdev2", 00:13:12.229 "uuid": "d94879de-263f-438d-8291-28c995ce7f39", 00:13:12.229 "is_configured": true, 00:13:12.229 "data_offset": 0, 00:13:12.229 "data_size": 65536 00:13:12.229 }, 00:13:12.229 { 00:13:12.229 "name": "BaseBdev3", 00:13:12.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.229 "is_configured": false, 00:13:12.229 "data_offset": 0, 00:13:12.229 "data_size": 0 00:13:12.229 } 00:13:12.229 ] 00:13:12.229 }' 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.229 14:12:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.799 [2024-09-30 14:12:17.173754] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:12.799 [2024-09-30 14:12:17.173896] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:13:12.799 [2024-09-30 14:12:17.173925] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:12.799 [2024-09-30 14:12:17.174256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:13:12.799 [2024-09-30 14:12:17.174717] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:13:12.799 [2024-09-30 14:12:17.174774] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:13:12.799 [2024-09-30 14:12:17.174987] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:12.799 BaseBdev3 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.799 [ 00:13:12.799 { 00:13:12.799 "name": "BaseBdev3", 00:13:12.799 "aliases": [ 00:13:12.799 "fafe75a1-5ea5-452c-bc8e-cbf2f9dad6b6" 00:13:12.799 ], 00:13:12.799 "product_name": "Malloc disk", 00:13:12.799 "block_size": 512, 00:13:12.799 "num_blocks": 65536, 00:13:12.799 "uuid": "fafe75a1-5ea5-452c-bc8e-cbf2f9dad6b6", 00:13:12.799 "assigned_rate_limits": { 00:13:12.799 "rw_ios_per_sec": 0, 00:13:12.799 "rw_mbytes_per_sec": 0, 00:13:12.799 "r_mbytes_per_sec": 0, 00:13:12.799 "w_mbytes_per_sec": 0 00:13:12.799 }, 00:13:12.799 "claimed": true, 00:13:12.799 "claim_type": "exclusive_write", 00:13:12.799 "zoned": false, 00:13:12.799 "supported_io_types": { 00:13:12.799 "read": true, 00:13:12.799 "write": true, 00:13:12.799 "unmap": true, 00:13:12.799 "flush": true, 00:13:12.799 "reset": true, 00:13:12.799 "nvme_admin": false, 00:13:12.799 "nvme_io": false, 00:13:12.799 "nvme_io_md": false, 00:13:12.799 "write_zeroes": true, 00:13:12.799 "zcopy": true, 00:13:12.799 "get_zone_info": false, 00:13:12.799 "zone_management": false, 00:13:12.799 "zone_append": false, 00:13:12.799 "compare": false, 00:13:12.799 "compare_and_write": false, 00:13:12.799 "abort": true, 00:13:12.799 "seek_hole": false, 00:13:12.799 "seek_data": false, 00:13:12.799 "copy": true, 00:13:12.799 "nvme_iov_md": false 00:13:12.799 }, 00:13:12.799 "memory_domains": [ 00:13:12.799 { 00:13:12.799 "dma_device_id": "system", 00:13:12.799 "dma_device_type": 1 00:13:12.799 }, 00:13:12.799 { 00:13:12.799 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:12.799 "dma_device_type": 2 00:13:12.799 } 00:13:12.799 ], 00:13:12.799 "driver_specific": {} 00:13:12.799 } 00:13:12.799 ] 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.799 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.800 "name": "Existed_Raid", 00:13:12.800 "uuid": "f9e1f3cd-8f39-4b8f-ab7a-4dd71f609fb1", 00:13:12.800 "strip_size_kb": 64, 00:13:12.800 "state": "online", 00:13:12.800 "raid_level": "raid5f", 00:13:12.800 "superblock": false, 00:13:12.800 "num_base_bdevs": 3, 00:13:12.800 "num_base_bdevs_discovered": 3, 00:13:12.800 "num_base_bdevs_operational": 3, 00:13:12.800 "base_bdevs_list": [ 00:13:12.800 { 00:13:12.800 "name": "BaseBdev1", 00:13:12.800 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:12.800 "is_configured": true, 00:13:12.800 "data_offset": 0, 00:13:12.800 "data_size": 65536 00:13:12.800 }, 00:13:12.800 { 00:13:12.800 "name": "BaseBdev2", 00:13:12.800 "uuid": "d94879de-263f-438d-8291-28c995ce7f39", 00:13:12.800 "is_configured": true, 00:13:12.800 "data_offset": 0, 00:13:12.800 "data_size": 65536 00:13:12.800 }, 00:13:12.800 { 00:13:12.800 "name": "BaseBdev3", 00:13:12.800 "uuid": "fafe75a1-5ea5-452c-bc8e-cbf2f9dad6b6", 00:13:12.800 "is_configured": true, 00:13:12.800 "data_offset": 0, 00:13:12.800 "data_size": 65536 00:13:12.800 } 00:13:12.800 ] 00:13:12.800 }' 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.800 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.060 [2024-09-30 14:12:17.677101] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:13.060 "name": "Existed_Raid", 00:13:13.060 "aliases": [ 00:13:13.060 "f9e1f3cd-8f39-4b8f-ab7a-4dd71f609fb1" 00:13:13.060 ], 00:13:13.060 "product_name": "Raid Volume", 00:13:13.060 "block_size": 512, 00:13:13.060 "num_blocks": 131072, 00:13:13.060 "uuid": "f9e1f3cd-8f39-4b8f-ab7a-4dd71f609fb1", 00:13:13.060 "assigned_rate_limits": { 00:13:13.060 "rw_ios_per_sec": 0, 00:13:13.060 "rw_mbytes_per_sec": 0, 00:13:13.060 "r_mbytes_per_sec": 0, 00:13:13.060 "w_mbytes_per_sec": 0 00:13:13.060 }, 00:13:13.060 "claimed": false, 00:13:13.060 "zoned": false, 00:13:13.060 "supported_io_types": { 00:13:13.060 "read": true, 00:13:13.060 "write": true, 00:13:13.060 "unmap": false, 00:13:13.060 "flush": false, 00:13:13.060 "reset": true, 00:13:13.060 "nvme_admin": false, 00:13:13.060 "nvme_io": false, 00:13:13.060 "nvme_io_md": false, 00:13:13.060 "write_zeroes": true, 00:13:13.060 "zcopy": false, 00:13:13.060 "get_zone_info": false, 00:13:13.060 "zone_management": false, 00:13:13.060 "zone_append": false, 00:13:13.060 "compare": false, 00:13:13.060 "compare_and_write": false, 00:13:13.060 "abort": false, 00:13:13.060 "seek_hole": false, 00:13:13.060 "seek_data": false, 00:13:13.060 "copy": false, 00:13:13.060 "nvme_iov_md": false 00:13:13.060 }, 00:13:13.060 "driver_specific": { 00:13:13.060 "raid": { 00:13:13.060 "uuid": "f9e1f3cd-8f39-4b8f-ab7a-4dd71f609fb1", 00:13:13.060 "strip_size_kb": 64, 00:13:13.060 "state": "online", 00:13:13.060 "raid_level": "raid5f", 00:13:13.060 "superblock": false, 00:13:13.060 "num_base_bdevs": 3, 00:13:13.060 "num_base_bdevs_discovered": 3, 00:13:13.060 "num_base_bdevs_operational": 3, 00:13:13.060 "base_bdevs_list": [ 00:13:13.060 { 00:13:13.060 "name": "BaseBdev1", 00:13:13.060 "uuid": "8ab6b0d9-3b38-4661-b9ce-4bc8493d2ada", 00:13:13.060 "is_configured": true, 00:13:13.060 "data_offset": 0, 00:13:13.060 "data_size": 65536 00:13:13.060 }, 00:13:13.060 { 00:13:13.060 "name": "BaseBdev2", 00:13:13.060 "uuid": "d94879de-263f-438d-8291-28c995ce7f39", 00:13:13.060 "is_configured": true, 00:13:13.060 "data_offset": 0, 00:13:13.060 "data_size": 65536 00:13:13.060 }, 00:13:13.060 { 00:13:13.060 "name": "BaseBdev3", 00:13:13.060 "uuid": "fafe75a1-5ea5-452c-bc8e-cbf2f9dad6b6", 00:13:13.060 "is_configured": true, 00:13:13.060 "data_offset": 0, 00:13:13.060 "data_size": 65536 00:13:13.060 } 00:13:13.060 ] 00:13:13.060 } 00:13:13.060 } 00:13:13.060 }' 00:13:13.060 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:13.320 BaseBdev2 00:13:13.320 BaseBdev3' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.320 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.321 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.321 [2024-09-30 14:12:17.972446] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.581 14:12:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.581 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.581 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.581 "name": "Existed_Raid", 00:13:13.581 "uuid": "f9e1f3cd-8f39-4b8f-ab7a-4dd71f609fb1", 00:13:13.581 "strip_size_kb": 64, 00:13:13.581 "state": "online", 00:13:13.581 "raid_level": "raid5f", 00:13:13.581 "superblock": false, 00:13:13.581 "num_base_bdevs": 3, 00:13:13.581 "num_base_bdevs_discovered": 2, 00:13:13.581 "num_base_bdevs_operational": 2, 00:13:13.581 "base_bdevs_list": [ 00:13:13.581 { 00:13:13.581 "name": null, 00:13:13.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.581 "is_configured": false, 00:13:13.581 "data_offset": 0, 00:13:13.581 "data_size": 65536 00:13:13.581 }, 00:13:13.581 { 00:13:13.581 "name": "BaseBdev2", 00:13:13.581 "uuid": "d94879de-263f-438d-8291-28c995ce7f39", 00:13:13.581 "is_configured": true, 00:13:13.581 "data_offset": 0, 00:13:13.581 "data_size": 65536 00:13:13.581 }, 00:13:13.581 { 00:13:13.581 "name": "BaseBdev3", 00:13:13.581 "uuid": "fafe75a1-5ea5-452c-bc8e-cbf2f9dad6b6", 00:13:13.581 "is_configured": true, 00:13:13.581 "data_offset": 0, 00:13:13.581 "data_size": 65536 00:13:13.581 } 00:13:13.581 ] 00:13:13.581 }' 00:13:13.581 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.581 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.841 [2024-09-30 14:12:18.483072] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:13.841 [2024-09-30 14:12:18.483168] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:13.841 [2024-09-30 14:12:18.493999] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:13.841 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 [2024-09-30 14:12:18.549921] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:14.106 [2024-09-30 14:12:18.550018] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 BaseBdev2 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 [ 00:13:14.106 { 00:13:14.106 "name": "BaseBdev2", 00:13:14.106 "aliases": [ 00:13:14.106 "78224bb0-171c-4a42-b096-f1dcfefda7a5" 00:13:14.106 ], 00:13:14.106 "product_name": "Malloc disk", 00:13:14.106 "block_size": 512, 00:13:14.106 "num_blocks": 65536, 00:13:14.106 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:14.106 "assigned_rate_limits": { 00:13:14.106 "rw_ios_per_sec": 0, 00:13:14.106 "rw_mbytes_per_sec": 0, 00:13:14.106 "r_mbytes_per_sec": 0, 00:13:14.106 "w_mbytes_per_sec": 0 00:13:14.106 }, 00:13:14.106 "claimed": false, 00:13:14.106 "zoned": false, 00:13:14.106 "supported_io_types": { 00:13:14.106 "read": true, 00:13:14.106 "write": true, 00:13:14.106 "unmap": true, 00:13:14.106 "flush": true, 00:13:14.106 "reset": true, 00:13:14.106 "nvme_admin": false, 00:13:14.106 "nvme_io": false, 00:13:14.106 "nvme_io_md": false, 00:13:14.106 "write_zeroes": true, 00:13:14.106 "zcopy": true, 00:13:14.106 "get_zone_info": false, 00:13:14.106 "zone_management": false, 00:13:14.106 "zone_append": false, 00:13:14.106 "compare": false, 00:13:14.106 "compare_and_write": false, 00:13:14.106 "abort": true, 00:13:14.106 "seek_hole": false, 00:13:14.106 "seek_data": false, 00:13:14.106 "copy": true, 00:13:14.106 "nvme_iov_md": false 00:13:14.106 }, 00:13:14.106 "memory_domains": [ 00:13:14.106 { 00:13:14.106 "dma_device_id": "system", 00:13:14.106 "dma_device_type": 1 00:13:14.106 }, 00:13:14.106 { 00:13:14.106 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.106 "dma_device_type": 2 00:13:14.106 } 00:13:14.106 ], 00:13:14.106 "driver_specific": {} 00:13:14.106 } 00:13:14.106 ] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 BaseBdev3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 [ 00:13:14.106 { 00:13:14.106 "name": "BaseBdev3", 00:13:14.106 "aliases": [ 00:13:14.106 "2c39e036-21be-461d-aadb-0bf85d88fd47" 00:13:14.106 ], 00:13:14.106 "product_name": "Malloc disk", 00:13:14.106 "block_size": 512, 00:13:14.106 "num_blocks": 65536, 00:13:14.106 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:14.106 "assigned_rate_limits": { 00:13:14.106 "rw_ios_per_sec": 0, 00:13:14.106 "rw_mbytes_per_sec": 0, 00:13:14.106 "r_mbytes_per_sec": 0, 00:13:14.106 "w_mbytes_per_sec": 0 00:13:14.106 }, 00:13:14.106 "claimed": false, 00:13:14.106 "zoned": false, 00:13:14.106 "supported_io_types": { 00:13:14.106 "read": true, 00:13:14.106 "write": true, 00:13:14.106 "unmap": true, 00:13:14.106 "flush": true, 00:13:14.106 "reset": true, 00:13:14.106 "nvme_admin": false, 00:13:14.106 "nvme_io": false, 00:13:14.106 "nvme_io_md": false, 00:13:14.106 "write_zeroes": true, 00:13:14.106 "zcopy": true, 00:13:14.106 "get_zone_info": false, 00:13:14.106 "zone_management": false, 00:13:14.106 "zone_append": false, 00:13:14.106 "compare": false, 00:13:14.106 "compare_and_write": false, 00:13:14.106 "abort": true, 00:13:14.106 "seek_hole": false, 00:13:14.106 "seek_data": false, 00:13:14.106 "copy": true, 00:13:14.106 "nvme_iov_md": false 00:13:14.106 }, 00:13:14.106 "memory_domains": [ 00:13:14.106 { 00:13:14.106 "dma_device_id": "system", 00:13:14.106 "dma_device_type": 1 00:13:14.106 }, 00:13:14.106 { 00:13:14.106 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.106 "dma_device_type": 2 00:13:14.106 } 00:13:14.106 ], 00:13:14.106 "driver_specific": {} 00:13:14.106 } 00:13:14.106 ] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 [2024-09-30 14:12:18.719444] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:14.106 [2024-09-30 14:12:18.719584] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:14.106 [2024-09-30 14:12:18.719624] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:14.106 [2024-09-30 14:12:18.721419] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.106 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.365 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.365 "name": "Existed_Raid", 00:13:14.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.365 "strip_size_kb": 64, 00:13:14.365 "state": "configuring", 00:13:14.365 "raid_level": "raid5f", 00:13:14.365 "superblock": false, 00:13:14.365 "num_base_bdevs": 3, 00:13:14.365 "num_base_bdevs_discovered": 2, 00:13:14.365 "num_base_bdevs_operational": 3, 00:13:14.365 "base_bdevs_list": [ 00:13:14.365 { 00:13:14.365 "name": "BaseBdev1", 00:13:14.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.365 "is_configured": false, 00:13:14.365 "data_offset": 0, 00:13:14.365 "data_size": 0 00:13:14.365 }, 00:13:14.365 { 00:13:14.365 "name": "BaseBdev2", 00:13:14.365 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:14.365 "is_configured": true, 00:13:14.365 "data_offset": 0, 00:13:14.365 "data_size": 65536 00:13:14.365 }, 00:13:14.365 { 00:13:14.365 "name": "BaseBdev3", 00:13:14.365 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:14.365 "is_configured": true, 00:13:14.365 "data_offset": 0, 00:13:14.365 "data_size": 65536 00:13:14.365 } 00:13:14.365 ] 00:13:14.365 }' 00:13:14.365 14:12:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.365 14:12:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.625 [2024-09-30 14:12:19.194622] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.625 "name": "Existed_Raid", 00:13:14.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.625 "strip_size_kb": 64, 00:13:14.625 "state": "configuring", 00:13:14.625 "raid_level": "raid5f", 00:13:14.625 "superblock": false, 00:13:14.625 "num_base_bdevs": 3, 00:13:14.625 "num_base_bdevs_discovered": 1, 00:13:14.625 "num_base_bdevs_operational": 3, 00:13:14.625 "base_bdevs_list": [ 00:13:14.625 { 00:13:14.625 "name": "BaseBdev1", 00:13:14.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.625 "is_configured": false, 00:13:14.625 "data_offset": 0, 00:13:14.625 "data_size": 0 00:13:14.625 }, 00:13:14.625 { 00:13:14.625 "name": null, 00:13:14.625 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:14.625 "is_configured": false, 00:13:14.625 "data_offset": 0, 00:13:14.625 "data_size": 65536 00:13:14.625 }, 00:13:14.625 { 00:13:14.625 "name": "BaseBdev3", 00:13:14.625 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:14.625 "is_configured": true, 00:13:14.625 "data_offset": 0, 00:13:14.625 "data_size": 65536 00:13:14.625 } 00:13:14.625 ] 00:13:14.625 }' 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.625 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.195 [2024-09-30 14:12:19.684783] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:15.195 BaseBdev1 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.195 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.195 [ 00:13:15.195 { 00:13:15.195 "name": "BaseBdev1", 00:13:15.195 "aliases": [ 00:13:15.195 "eb4eca87-8609-4d45-a8ca-500ba1ffd71e" 00:13:15.195 ], 00:13:15.195 "product_name": "Malloc disk", 00:13:15.195 "block_size": 512, 00:13:15.195 "num_blocks": 65536, 00:13:15.195 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:15.195 "assigned_rate_limits": { 00:13:15.195 "rw_ios_per_sec": 0, 00:13:15.195 "rw_mbytes_per_sec": 0, 00:13:15.195 "r_mbytes_per_sec": 0, 00:13:15.195 "w_mbytes_per_sec": 0 00:13:15.195 }, 00:13:15.195 "claimed": true, 00:13:15.195 "claim_type": "exclusive_write", 00:13:15.195 "zoned": false, 00:13:15.195 "supported_io_types": { 00:13:15.195 "read": true, 00:13:15.195 "write": true, 00:13:15.195 "unmap": true, 00:13:15.195 "flush": true, 00:13:15.195 "reset": true, 00:13:15.195 "nvme_admin": false, 00:13:15.195 "nvme_io": false, 00:13:15.195 "nvme_io_md": false, 00:13:15.195 "write_zeroes": true, 00:13:15.195 "zcopy": true, 00:13:15.196 "get_zone_info": false, 00:13:15.196 "zone_management": false, 00:13:15.196 "zone_append": false, 00:13:15.196 "compare": false, 00:13:15.196 "compare_and_write": false, 00:13:15.196 "abort": true, 00:13:15.196 "seek_hole": false, 00:13:15.196 "seek_data": false, 00:13:15.196 "copy": true, 00:13:15.196 "nvme_iov_md": false 00:13:15.196 }, 00:13:15.196 "memory_domains": [ 00:13:15.196 { 00:13:15.196 "dma_device_id": "system", 00:13:15.196 "dma_device_type": 1 00:13:15.196 }, 00:13:15.196 { 00:13:15.196 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:15.196 "dma_device_type": 2 00:13:15.196 } 00:13:15.196 ], 00:13:15.196 "driver_specific": {} 00:13:15.196 } 00:13:15.196 ] 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.196 "name": "Existed_Raid", 00:13:15.196 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.196 "strip_size_kb": 64, 00:13:15.196 "state": "configuring", 00:13:15.196 "raid_level": "raid5f", 00:13:15.196 "superblock": false, 00:13:15.196 "num_base_bdevs": 3, 00:13:15.196 "num_base_bdevs_discovered": 2, 00:13:15.196 "num_base_bdevs_operational": 3, 00:13:15.196 "base_bdevs_list": [ 00:13:15.196 { 00:13:15.196 "name": "BaseBdev1", 00:13:15.196 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:15.196 "is_configured": true, 00:13:15.196 "data_offset": 0, 00:13:15.196 "data_size": 65536 00:13:15.196 }, 00:13:15.196 { 00:13:15.196 "name": null, 00:13:15.196 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:15.196 "is_configured": false, 00:13:15.196 "data_offset": 0, 00:13:15.196 "data_size": 65536 00:13:15.196 }, 00:13:15.196 { 00:13:15.196 "name": "BaseBdev3", 00:13:15.196 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:15.196 "is_configured": true, 00:13:15.196 "data_offset": 0, 00:13:15.196 "data_size": 65536 00:13:15.196 } 00:13:15.196 ] 00:13:15.196 }' 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.196 14:12:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.765 [2024-09-30 14:12:20.235913] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:15.765 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.766 "name": "Existed_Raid", 00:13:15.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.766 "strip_size_kb": 64, 00:13:15.766 "state": "configuring", 00:13:15.766 "raid_level": "raid5f", 00:13:15.766 "superblock": false, 00:13:15.766 "num_base_bdevs": 3, 00:13:15.766 "num_base_bdevs_discovered": 1, 00:13:15.766 "num_base_bdevs_operational": 3, 00:13:15.766 "base_bdevs_list": [ 00:13:15.766 { 00:13:15.766 "name": "BaseBdev1", 00:13:15.766 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:15.766 "is_configured": true, 00:13:15.766 "data_offset": 0, 00:13:15.766 "data_size": 65536 00:13:15.766 }, 00:13:15.766 { 00:13:15.766 "name": null, 00:13:15.766 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:15.766 "is_configured": false, 00:13:15.766 "data_offset": 0, 00:13:15.766 "data_size": 65536 00:13:15.766 }, 00:13:15.766 { 00:13:15.766 "name": null, 00:13:15.766 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:15.766 "is_configured": false, 00:13:15.766 "data_offset": 0, 00:13:15.766 "data_size": 65536 00:13:15.766 } 00:13:15.766 ] 00:13:15.766 }' 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.766 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.026 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.026 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.026 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.026 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:16.026 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.286 [2024-09-30 14:12:20.691262] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.286 "name": "Existed_Raid", 00:13:16.286 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.286 "strip_size_kb": 64, 00:13:16.286 "state": "configuring", 00:13:16.286 "raid_level": "raid5f", 00:13:16.286 "superblock": false, 00:13:16.286 "num_base_bdevs": 3, 00:13:16.286 "num_base_bdevs_discovered": 2, 00:13:16.286 "num_base_bdevs_operational": 3, 00:13:16.286 "base_bdevs_list": [ 00:13:16.286 { 00:13:16.286 "name": "BaseBdev1", 00:13:16.286 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:16.286 "is_configured": true, 00:13:16.286 "data_offset": 0, 00:13:16.286 "data_size": 65536 00:13:16.286 }, 00:13:16.286 { 00:13:16.286 "name": null, 00:13:16.286 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:16.286 "is_configured": false, 00:13:16.286 "data_offset": 0, 00:13:16.286 "data_size": 65536 00:13:16.286 }, 00:13:16.286 { 00:13:16.286 "name": "BaseBdev3", 00:13:16.286 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:16.286 "is_configured": true, 00:13:16.286 "data_offset": 0, 00:13:16.286 "data_size": 65536 00:13:16.286 } 00:13:16.286 ] 00:13:16.286 }' 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.286 14:12:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.546 [2024-09-30 14:12:21.174392] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.546 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:16.806 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.806 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.806 "name": "Existed_Raid", 00:13:16.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.806 "strip_size_kb": 64, 00:13:16.806 "state": "configuring", 00:13:16.806 "raid_level": "raid5f", 00:13:16.806 "superblock": false, 00:13:16.806 "num_base_bdevs": 3, 00:13:16.806 "num_base_bdevs_discovered": 1, 00:13:16.806 "num_base_bdevs_operational": 3, 00:13:16.806 "base_bdevs_list": [ 00:13:16.806 { 00:13:16.806 "name": null, 00:13:16.806 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:16.806 "is_configured": false, 00:13:16.806 "data_offset": 0, 00:13:16.806 "data_size": 65536 00:13:16.806 }, 00:13:16.806 { 00:13:16.806 "name": null, 00:13:16.806 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:16.806 "is_configured": false, 00:13:16.806 "data_offset": 0, 00:13:16.806 "data_size": 65536 00:13:16.806 }, 00:13:16.806 { 00:13:16.806 "name": "BaseBdev3", 00:13:16.806 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:16.806 "is_configured": true, 00:13:16.806 "data_offset": 0, 00:13:16.806 "data_size": 65536 00:13:16.806 } 00:13:16.806 ] 00:13:16.806 }' 00:13:16.806 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.806 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.066 [2024-09-30 14:12:21.632158] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.066 "name": "Existed_Raid", 00:13:17.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.066 "strip_size_kb": 64, 00:13:17.066 "state": "configuring", 00:13:17.066 "raid_level": "raid5f", 00:13:17.066 "superblock": false, 00:13:17.066 "num_base_bdevs": 3, 00:13:17.066 "num_base_bdevs_discovered": 2, 00:13:17.066 "num_base_bdevs_operational": 3, 00:13:17.066 "base_bdevs_list": [ 00:13:17.066 { 00:13:17.066 "name": null, 00:13:17.066 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:17.066 "is_configured": false, 00:13:17.066 "data_offset": 0, 00:13:17.066 "data_size": 65536 00:13:17.066 }, 00:13:17.066 { 00:13:17.066 "name": "BaseBdev2", 00:13:17.066 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:17.066 "is_configured": true, 00:13:17.066 "data_offset": 0, 00:13:17.066 "data_size": 65536 00:13:17.066 }, 00:13:17.066 { 00:13:17.066 "name": "BaseBdev3", 00:13:17.066 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:17.066 "is_configured": true, 00:13:17.066 "data_offset": 0, 00:13:17.066 "data_size": 65536 00:13:17.066 } 00:13:17.066 ] 00:13:17.066 }' 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.066 14:12:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u eb4eca87-8609-4d45-a8ca-500ba1ffd71e 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 [2024-09-30 14:12:22.133895] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:17.644 [2024-09-30 14:12:22.133997] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:13:17.644 [2024-09-30 14:12:22.134022] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:17.644 [2024-09-30 14:12:22.134274] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:13:17.644 [2024-09-30 14:12:22.134706] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:13:17.644 [2024-09-30 14:12:22.134758] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:13:17.644 [2024-09-30 14:12:22.134948] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:17.644 NewBaseBdev 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.644 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.644 [ 00:13:17.644 { 00:13:17.644 "name": "NewBaseBdev", 00:13:17.644 "aliases": [ 00:13:17.644 "eb4eca87-8609-4d45-a8ca-500ba1ffd71e" 00:13:17.644 ], 00:13:17.644 "product_name": "Malloc disk", 00:13:17.644 "block_size": 512, 00:13:17.644 "num_blocks": 65536, 00:13:17.644 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:17.644 "assigned_rate_limits": { 00:13:17.644 "rw_ios_per_sec": 0, 00:13:17.644 "rw_mbytes_per_sec": 0, 00:13:17.644 "r_mbytes_per_sec": 0, 00:13:17.644 "w_mbytes_per_sec": 0 00:13:17.644 }, 00:13:17.644 "claimed": true, 00:13:17.644 "claim_type": "exclusive_write", 00:13:17.644 "zoned": false, 00:13:17.644 "supported_io_types": { 00:13:17.644 "read": true, 00:13:17.644 "write": true, 00:13:17.644 "unmap": true, 00:13:17.644 "flush": true, 00:13:17.644 "reset": true, 00:13:17.644 "nvme_admin": false, 00:13:17.644 "nvme_io": false, 00:13:17.644 "nvme_io_md": false, 00:13:17.644 "write_zeroes": true, 00:13:17.644 "zcopy": true, 00:13:17.644 "get_zone_info": false, 00:13:17.644 "zone_management": false, 00:13:17.644 "zone_append": false, 00:13:17.644 "compare": false, 00:13:17.644 "compare_and_write": false, 00:13:17.644 "abort": true, 00:13:17.644 "seek_hole": false, 00:13:17.644 "seek_data": false, 00:13:17.644 "copy": true, 00:13:17.644 "nvme_iov_md": false 00:13:17.644 }, 00:13:17.644 "memory_domains": [ 00:13:17.644 { 00:13:17.644 "dma_device_id": "system", 00:13:17.644 "dma_device_type": 1 00:13:17.644 }, 00:13:17.644 { 00:13:17.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:17.644 "dma_device_type": 2 00:13:17.644 } 00:13:17.644 ], 00:13:17.644 "driver_specific": {} 00:13:17.644 } 00:13:17.644 ] 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.645 "name": "Existed_Raid", 00:13:17.645 "uuid": "ef413880-c64b-4991-9273-3c8d1e26e22b", 00:13:17.645 "strip_size_kb": 64, 00:13:17.645 "state": "online", 00:13:17.645 "raid_level": "raid5f", 00:13:17.645 "superblock": false, 00:13:17.645 "num_base_bdevs": 3, 00:13:17.645 "num_base_bdevs_discovered": 3, 00:13:17.645 "num_base_bdevs_operational": 3, 00:13:17.645 "base_bdevs_list": [ 00:13:17.645 { 00:13:17.645 "name": "NewBaseBdev", 00:13:17.645 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:17.645 "is_configured": true, 00:13:17.645 "data_offset": 0, 00:13:17.645 "data_size": 65536 00:13:17.645 }, 00:13:17.645 { 00:13:17.645 "name": "BaseBdev2", 00:13:17.645 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:17.645 "is_configured": true, 00:13:17.645 "data_offset": 0, 00:13:17.645 "data_size": 65536 00:13:17.645 }, 00:13:17.645 { 00:13:17.645 "name": "BaseBdev3", 00:13:17.645 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:17.645 "is_configured": true, 00:13:17.645 "data_offset": 0, 00:13:17.645 "data_size": 65536 00:13:17.645 } 00:13:17.645 ] 00:13:17.645 }' 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.645 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:18.217 [2024-09-30 14:12:22.649193] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.217 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:18.217 "name": "Existed_Raid", 00:13:18.217 "aliases": [ 00:13:18.217 "ef413880-c64b-4991-9273-3c8d1e26e22b" 00:13:18.217 ], 00:13:18.217 "product_name": "Raid Volume", 00:13:18.217 "block_size": 512, 00:13:18.217 "num_blocks": 131072, 00:13:18.217 "uuid": "ef413880-c64b-4991-9273-3c8d1e26e22b", 00:13:18.217 "assigned_rate_limits": { 00:13:18.217 "rw_ios_per_sec": 0, 00:13:18.217 "rw_mbytes_per_sec": 0, 00:13:18.217 "r_mbytes_per_sec": 0, 00:13:18.217 "w_mbytes_per_sec": 0 00:13:18.217 }, 00:13:18.217 "claimed": false, 00:13:18.217 "zoned": false, 00:13:18.217 "supported_io_types": { 00:13:18.217 "read": true, 00:13:18.217 "write": true, 00:13:18.217 "unmap": false, 00:13:18.217 "flush": false, 00:13:18.217 "reset": true, 00:13:18.217 "nvme_admin": false, 00:13:18.217 "nvme_io": false, 00:13:18.217 "nvme_io_md": false, 00:13:18.217 "write_zeroes": true, 00:13:18.217 "zcopy": false, 00:13:18.217 "get_zone_info": false, 00:13:18.217 "zone_management": false, 00:13:18.217 "zone_append": false, 00:13:18.217 "compare": false, 00:13:18.217 "compare_and_write": false, 00:13:18.217 "abort": false, 00:13:18.217 "seek_hole": false, 00:13:18.217 "seek_data": false, 00:13:18.217 "copy": false, 00:13:18.217 "nvme_iov_md": false 00:13:18.217 }, 00:13:18.217 "driver_specific": { 00:13:18.217 "raid": { 00:13:18.217 "uuid": "ef413880-c64b-4991-9273-3c8d1e26e22b", 00:13:18.217 "strip_size_kb": 64, 00:13:18.217 "state": "online", 00:13:18.217 "raid_level": "raid5f", 00:13:18.217 "superblock": false, 00:13:18.217 "num_base_bdevs": 3, 00:13:18.217 "num_base_bdevs_discovered": 3, 00:13:18.217 "num_base_bdevs_operational": 3, 00:13:18.217 "base_bdevs_list": [ 00:13:18.217 { 00:13:18.218 "name": "NewBaseBdev", 00:13:18.218 "uuid": "eb4eca87-8609-4d45-a8ca-500ba1ffd71e", 00:13:18.218 "is_configured": true, 00:13:18.218 "data_offset": 0, 00:13:18.218 "data_size": 65536 00:13:18.218 }, 00:13:18.218 { 00:13:18.218 "name": "BaseBdev2", 00:13:18.218 "uuid": "78224bb0-171c-4a42-b096-f1dcfefda7a5", 00:13:18.218 "is_configured": true, 00:13:18.218 "data_offset": 0, 00:13:18.218 "data_size": 65536 00:13:18.218 }, 00:13:18.218 { 00:13:18.218 "name": "BaseBdev3", 00:13:18.218 "uuid": "2c39e036-21be-461d-aadb-0bf85d88fd47", 00:13:18.218 "is_configured": true, 00:13:18.218 "data_offset": 0, 00:13:18.218 "data_size": 65536 00:13:18.218 } 00:13:18.218 ] 00:13:18.218 } 00:13:18.218 } 00:13:18.218 }' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:18.218 BaseBdev2 00:13:18.218 BaseBdev3' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.218 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.478 [2024-09-30 14:12:22.912569] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:18.478 [2024-09-30 14:12:22.912636] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:18.478 [2024-09-30 14:12:22.912705] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:18.478 [2024-09-30 14:12:22.912941] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:18.478 [2024-09-30 14:12:22.913011] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 91474 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 91474 ']' 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 91474 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91474 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91474' 00:13:18.478 killing process with pid 91474 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 91474 00:13:18.478 [2024-09-30 14:12:22.966526] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:18.478 14:12:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 91474 00:13:18.478 [2024-09-30 14:12:22.996204] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:18.739 00:13:18.739 real 0m8.868s 00:13:18.739 user 0m15.077s 00:13:18.739 sys 0m1.905s 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.739 ************************************ 00:13:18.739 END TEST raid5f_state_function_test 00:13:18.739 ************************************ 00:13:18.739 14:12:23 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:18.739 14:12:23 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:18.739 14:12:23 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:18.739 14:12:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:18.739 ************************************ 00:13:18.739 START TEST raid5f_state_function_test_sb 00:13:18.739 ************************************ 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 true 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=92079 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:18.739 Process raid pid: 92079 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 92079' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 92079 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 92079 ']' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:18.739 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:18.739 14:12:23 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.000 [2024-09-30 14:12:23.416538] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:13:19.000 [2024-09-30 14:12:23.416748] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:19.000 [2024-09-30 14:12:23.549697] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:13:19.000 [2024-09-30 14:12:23.578474] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:19.000 [2024-09-30 14:12:23.623401] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:19.260 [2024-09-30 14:12:23.665529] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:19.260 [2024-09-30 14:12:23.665563] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.830 [2024-09-30 14:12:24.230695] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:19.830 [2024-09-30 14:12:24.230818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:19.830 [2024-09-30 14:12:24.230837] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:19.830 [2024-09-30 14:12:24.230845] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:19.830 [2024-09-30 14:12:24.230855] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:19.830 [2024-09-30 14:12:24.230861] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.830 "name": "Existed_Raid", 00:13:19.830 "uuid": "5193c253-bde6-4ae6-8411-233e98613b12", 00:13:19.830 "strip_size_kb": 64, 00:13:19.830 "state": "configuring", 00:13:19.830 "raid_level": "raid5f", 00:13:19.830 "superblock": true, 00:13:19.830 "num_base_bdevs": 3, 00:13:19.830 "num_base_bdevs_discovered": 0, 00:13:19.830 "num_base_bdevs_operational": 3, 00:13:19.830 "base_bdevs_list": [ 00:13:19.830 { 00:13:19.830 "name": "BaseBdev1", 00:13:19.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.830 "is_configured": false, 00:13:19.830 "data_offset": 0, 00:13:19.830 "data_size": 0 00:13:19.830 }, 00:13:19.830 { 00:13:19.830 "name": "BaseBdev2", 00:13:19.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.830 "is_configured": false, 00:13:19.830 "data_offset": 0, 00:13:19.830 "data_size": 0 00:13:19.830 }, 00:13:19.830 { 00:13:19.830 "name": "BaseBdev3", 00:13:19.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.830 "is_configured": false, 00:13:19.830 "data_offset": 0, 00:13:19.830 "data_size": 0 00:13:19.830 } 00:13:19.830 ] 00:13:19.830 }' 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.830 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.090 [2024-09-30 14:12:24.717728] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:20.090 [2024-09-30 14:12:24.717804] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.090 [2024-09-30 14:12:24.729719] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:20.090 [2024-09-30 14:12:24.729788] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:20.090 [2024-09-30 14:12:24.729814] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:20.090 [2024-09-30 14:12:24.729833] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:20.090 [2024-09-30 14:12:24.729851] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:20.090 [2024-09-30 14:12:24.729868] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.090 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.350 [2024-09-30 14:12:24.750747] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:20.350 BaseBdev1 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.350 [ 00:13:20.350 { 00:13:20.350 "name": "BaseBdev1", 00:13:20.350 "aliases": [ 00:13:20.350 "d8de5e74-136a-4990-b777-343c6aaf660a" 00:13:20.350 ], 00:13:20.350 "product_name": "Malloc disk", 00:13:20.350 "block_size": 512, 00:13:20.350 "num_blocks": 65536, 00:13:20.350 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:20.350 "assigned_rate_limits": { 00:13:20.350 "rw_ios_per_sec": 0, 00:13:20.350 "rw_mbytes_per_sec": 0, 00:13:20.350 "r_mbytes_per_sec": 0, 00:13:20.350 "w_mbytes_per_sec": 0 00:13:20.350 }, 00:13:20.350 "claimed": true, 00:13:20.350 "claim_type": "exclusive_write", 00:13:20.350 "zoned": false, 00:13:20.350 "supported_io_types": { 00:13:20.350 "read": true, 00:13:20.350 "write": true, 00:13:20.350 "unmap": true, 00:13:20.350 "flush": true, 00:13:20.350 "reset": true, 00:13:20.350 "nvme_admin": false, 00:13:20.350 "nvme_io": false, 00:13:20.350 "nvme_io_md": false, 00:13:20.350 "write_zeroes": true, 00:13:20.350 "zcopy": true, 00:13:20.350 "get_zone_info": false, 00:13:20.350 "zone_management": false, 00:13:20.350 "zone_append": false, 00:13:20.350 "compare": false, 00:13:20.350 "compare_and_write": false, 00:13:20.350 "abort": true, 00:13:20.350 "seek_hole": false, 00:13:20.350 "seek_data": false, 00:13:20.350 "copy": true, 00:13:20.350 "nvme_iov_md": false 00:13:20.350 }, 00:13:20.350 "memory_domains": [ 00:13:20.350 { 00:13:20.350 "dma_device_id": "system", 00:13:20.350 "dma_device_type": 1 00:13:20.350 }, 00:13:20.350 { 00:13:20.350 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:20.350 "dma_device_type": 2 00:13:20.350 } 00:13:20.350 ], 00:13:20.350 "driver_specific": {} 00:13:20.350 } 00:13:20.350 ] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.350 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.350 "name": "Existed_Raid", 00:13:20.350 "uuid": "b3074069-3617-44d1-b171-c2ed31a4ce17", 00:13:20.350 "strip_size_kb": 64, 00:13:20.350 "state": "configuring", 00:13:20.350 "raid_level": "raid5f", 00:13:20.350 "superblock": true, 00:13:20.350 "num_base_bdevs": 3, 00:13:20.350 "num_base_bdevs_discovered": 1, 00:13:20.350 "num_base_bdevs_operational": 3, 00:13:20.350 "base_bdevs_list": [ 00:13:20.350 { 00:13:20.350 "name": "BaseBdev1", 00:13:20.350 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:20.350 "is_configured": true, 00:13:20.350 "data_offset": 2048, 00:13:20.350 "data_size": 63488 00:13:20.350 }, 00:13:20.350 { 00:13:20.350 "name": "BaseBdev2", 00:13:20.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.351 "is_configured": false, 00:13:20.351 "data_offset": 0, 00:13:20.351 "data_size": 0 00:13:20.351 }, 00:13:20.351 { 00:13:20.351 "name": "BaseBdev3", 00:13:20.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.351 "is_configured": false, 00:13:20.351 "data_offset": 0, 00:13:20.351 "data_size": 0 00:13:20.351 } 00:13:20.351 ] 00:13:20.351 }' 00:13:20.351 14:12:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.351 14:12:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.611 [2024-09-30 14:12:25.218062] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:20.611 [2024-09-30 14:12:25.218110] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.611 [2024-09-30 14:12:25.230064] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:20.611 [2024-09-30 14:12:25.231784] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:20.611 [2024-09-30 14:12:25.231819] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:20.611 [2024-09-30 14:12:25.231831] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:20.611 [2024-09-30 14:12:25.231837] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.611 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.870 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.870 "name": "Existed_Raid", 00:13:20.870 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:20.870 "strip_size_kb": 64, 00:13:20.870 "state": "configuring", 00:13:20.870 "raid_level": "raid5f", 00:13:20.871 "superblock": true, 00:13:20.871 "num_base_bdevs": 3, 00:13:20.871 "num_base_bdevs_discovered": 1, 00:13:20.871 "num_base_bdevs_operational": 3, 00:13:20.871 "base_bdevs_list": [ 00:13:20.871 { 00:13:20.871 "name": "BaseBdev1", 00:13:20.871 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:20.871 "is_configured": true, 00:13:20.871 "data_offset": 2048, 00:13:20.871 "data_size": 63488 00:13:20.871 }, 00:13:20.871 { 00:13:20.871 "name": "BaseBdev2", 00:13:20.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.871 "is_configured": false, 00:13:20.871 "data_offset": 0, 00:13:20.871 "data_size": 0 00:13:20.871 }, 00:13:20.871 { 00:13:20.871 "name": "BaseBdev3", 00:13:20.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.871 "is_configured": false, 00:13:20.871 "data_offset": 0, 00:13:20.871 "data_size": 0 00:13:20.871 } 00:13:20.871 ] 00:13:20.871 }' 00:13:20.871 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.871 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.131 [2024-09-30 14:12:25.717948] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:21.131 BaseBdev2 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.131 [ 00:13:21.131 { 00:13:21.131 "name": "BaseBdev2", 00:13:21.131 "aliases": [ 00:13:21.131 "5b5d431a-4f7d-49c5-b491-e91160973b8a" 00:13:21.131 ], 00:13:21.131 "product_name": "Malloc disk", 00:13:21.131 "block_size": 512, 00:13:21.131 "num_blocks": 65536, 00:13:21.131 "uuid": "5b5d431a-4f7d-49c5-b491-e91160973b8a", 00:13:21.131 "assigned_rate_limits": { 00:13:21.131 "rw_ios_per_sec": 0, 00:13:21.131 "rw_mbytes_per_sec": 0, 00:13:21.131 "r_mbytes_per_sec": 0, 00:13:21.131 "w_mbytes_per_sec": 0 00:13:21.131 }, 00:13:21.131 "claimed": true, 00:13:21.131 "claim_type": "exclusive_write", 00:13:21.131 "zoned": false, 00:13:21.131 "supported_io_types": { 00:13:21.131 "read": true, 00:13:21.131 "write": true, 00:13:21.131 "unmap": true, 00:13:21.131 "flush": true, 00:13:21.131 "reset": true, 00:13:21.131 "nvme_admin": false, 00:13:21.131 "nvme_io": false, 00:13:21.131 "nvme_io_md": false, 00:13:21.131 "write_zeroes": true, 00:13:21.131 "zcopy": true, 00:13:21.131 "get_zone_info": false, 00:13:21.131 "zone_management": false, 00:13:21.131 "zone_append": false, 00:13:21.131 "compare": false, 00:13:21.131 "compare_and_write": false, 00:13:21.131 "abort": true, 00:13:21.131 "seek_hole": false, 00:13:21.131 "seek_data": false, 00:13:21.131 "copy": true, 00:13:21.131 "nvme_iov_md": false 00:13:21.131 }, 00:13:21.131 "memory_domains": [ 00:13:21.131 { 00:13:21.131 "dma_device_id": "system", 00:13:21.131 "dma_device_type": 1 00:13:21.131 }, 00:13:21.131 { 00:13:21.131 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:21.131 "dma_device_type": 2 00:13:21.131 } 00:13:21.131 ], 00:13:21.131 "driver_specific": {} 00:13:21.131 } 00:13:21.131 ] 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.131 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.390 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.391 "name": "Existed_Raid", 00:13:21.391 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:21.391 "strip_size_kb": 64, 00:13:21.391 "state": "configuring", 00:13:21.391 "raid_level": "raid5f", 00:13:21.391 "superblock": true, 00:13:21.391 "num_base_bdevs": 3, 00:13:21.391 "num_base_bdevs_discovered": 2, 00:13:21.391 "num_base_bdevs_operational": 3, 00:13:21.391 "base_bdevs_list": [ 00:13:21.391 { 00:13:21.391 "name": "BaseBdev1", 00:13:21.391 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:21.391 "is_configured": true, 00:13:21.391 "data_offset": 2048, 00:13:21.391 "data_size": 63488 00:13:21.391 }, 00:13:21.391 { 00:13:21.391 "name": "BaseBdev2", 00:13:21.391 "uuid": "5b5d431a-4f7d-49c5-b491-e91160973b8a", 00:13:21.391 "is_configured": true, 00:13:21.391 "data_offset": 2048, 00:13:21.391 "data_size": 63488 00:13:21.391 }, 00:13:21.391 { 00:13:21.391 "name": "BaseBdev3", 00:13:21.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.391 "is_configured": false, 00:13:21.391 "data_offset": 0, 00:13:21.391 "data_size": 0 00:13:21.391 } 00:13:21.391 ] 00:13:21.391 }' 00:13:21.391 14:12:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.391 14:12:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.651 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:21.651 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.651 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.651 [2024-09-30 14:12:26.223982] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:21.652 [2024-09-30 14:12:26.224177] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:13:21.652 [2024-09-30 14:12:26.224191] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:21.652 [2024-09-30 14:12:26.224460] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:13:21.652 BaseBdev3 00:13:21.652 [2024-09-30 14:12:26.224894] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:13:21.652 [2024-09-30 14:12:26.224922] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:13:21.652 [2024-09-30 14:12:26.225031] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.652 [ 00:13:21.652 { 00:13:21.652 "name": "BaseBdev3", 00:13:21.652 "aliases": [ 00:13:21.652 "da5b5ffe-b96f-4adc-b06f-ce7f3284720f" 00:13:21.652 ], 00:13:21.652 "product_name": "Malloc disk", 00:13:21.652 "block_size": 512, 00:13:21.652 "num_blocks": 65536, 00:13:21.652 "uuid": "da5b5ffe-b96f-4adc-b06f-ce7f3284720f", 00:13:21.652 "assigned_rate_limits": { 00:13:21.652 "rw_ios_per_sec": 0, 00:13:21.652 "rw_mbytes_per_sec": 0, 00:13:21.652 "r_mbytes_per_sec": 0, 00:13:21.652 "w_mbytes_per_sec": 0 00:13:21.652 }, 00:13:21.652 "claimed": true, 00:13:21.652 "claim_type": "exclusive_write", 00:13:21.652 "zoned": false, 00:13:21.652 "supported_io_types": { 00:13:21.652 "read": true, 00:13:21.652 "write": true, 00:13:21.652 "unmap": true, 00:13:21.652 "flush": true, 00:13:21.652 "reset": true, 00:13:21.652 "nvme_admin": false, 00:13:21.652 "nvme_io": false, 00:13:21.652 "nvme_io_md": false, 00:13:21.652 "write_zeroes": true, 00:13:21.652 "zcopy": true, 00:13:21.652 "get_zone_info": false, 00:13:21.652 "zone_management": false, 00:13:21.652 "zone_append": false, 00:13:21.652 "compare": false, 00:13:21.652 "compare_and_write": false, 00:13:21.652 "abort": true, 00:13:21.652 "seek_hole": false, 00:13:21.652 "seek_data": false, 00:13:21.652 "copy": true, 00:13:21.652 "nvme_iov_md": false 00:13:21.652 }, 00:13:21.652 "memory_domains": [ 00:13:21.652 { 00:13:21.652 "dma_device_id": "system", 00:13:21.652 "dma_device_type": 1 00:13:21.652 }, 00:13:21.652 { 00:13:21.652 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:21.652 "dma_device_type": 2 00:13:21.652 } 00:13:21.652 ], 00:13:21.652 "driver_specific": {} 00:13:21.652 } 00:13:21.652 ] 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.652 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.912 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.912 "name": "Existed_Raid", 00:13:21.912 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:21.912 "strip_size_kb": 64, 00:13:21.912 "state": "online", 00:13:21.912 "raid_level": "raid5f", 00:13:21.912 "superblock": true, 00:13:21.912 "num_base_bdevs": 3, 00:13:21.912 "num_base_bdevs_discovered": 3, 00:13:21.912 "num_base_bdevs_operational": 3, 00:13:21.912 "base_bdevs_list": [ 00:13:21.912 { 00:13:21.912 "name": "BaseBdev1", 00:13:21.912 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:21.912 "is_configured": true, 00:13:21.912 "data_offset": 2048, 00:13:21.912 "data_size": 63488 00:13:21.912 }, 00:13:21.912 { 00:13:21.912 "name": "BaseBdev2", 00:13:21.912 "uuid": "5b5d431a-4f7d-49c5-b491-e91160973b8a", 00:13:21.912 "is_configured": true, 00:13:21.912 "data_offset": 2048, 00:13:21.912 "data_size": 63488 00:13:21.912 }, 00:13:21.912 { 00:13:21.912 "name": "BaseBdev3", 00:13:21.912 "uuid": "da5b5ffe-b96f-4adc-b06f-ce7f3284720f", 00:13:21.912 "is_configured": true, 00:13:21.912 "data_offset": 2048, 00:13:21.912 "data_size": 63488 00:13:21.912 } 00:13:21.912 ] 00:13:21.912 }' 00:13:21.912 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.912 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.173 [2024-09-30 14:12:26.747340] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:22.173 "name": "Existed_Raid", 00:13:22.173 "aliases": [ 00:13:22.173 "2d914aa2-2c80-493a-8233-742c84d8df28" 00:13:22.173 ], 00:13:22.173 "product_name": "Raid Volume", 00:13:22.173 "block_size": 512, 00:13:22.173 "num_blocks": 126976, 00:13:22.173 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:22.173 "assigned_rate_limits": { 00:13:22.173 "rw_ios_per_sec": 0, 00:13:22.173 "rw_mbytes_per_sec": 0, 00:13:22.173 "r_mbytes_per_sec": 0, 00:13:22.173 "w_mbytes_per_sec": 0 00:13:22.173 }, 00:13:22.173 "claimed": false, 00:13:22.173 "zoned": false, 00:13:22.173 "supported_io_types": { 00:13:22.173 "read": true, 00:13:22.173 "write": true, 00:13:22.173 "unmap": false, 00:13:22.173 "flush": false, 00:13:22.173 "reset": true, 00:13:22.173 "nvme_admin": false, 00:13:22.173 "nvme_io": false, 00:13:22.173 "nvme_io_md": false, 00:13:22.173 "write_zeroes": true, 00:13:22.173 "zcopy": false, 00:13:22.173 "get_zone_info": false, 00:13:22.173 "zone_management": false, 00:13:22.173 "zone_append": false, 00:13:22.173 "compare": false, 00:13:22.173 "compare_and_write": false, 00:13:22.173 "abort": false, 00:13:22.173 "seek_hole": false, 00:13:22.173 "seek_data": false, 00:13:22.173 "copy": false, 00:13:22.173 "nvme_iov_md": false 00:13:22.173 }, 00:13:22.173 "driver_specific": { 00:13:22.173 "raid": { 00:13:22.173 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:22.173 "strip_size_kb": 64, 00:13:22.173 "state": "online", 00:13:22.173 "raid_level": "raid5f", 00:13:22.173 "superblock": true, 00:13:22.173 "num_base_bdevs": 3, 00:13:22.173 "num_base_bdevs_discovered": 3, 00:13:22.173 "num_base_bdevs_operational": 3, 00:13:22.173 "base_bdevs_list": [ 00:13:22.173 { 00:13:22.173 "name": "BaseBdev1", 00:13:22.173 "uuid": "d8de5e74-136a-4990-b777-343c6aaf660a", 00:13:22.173 "is_configured": true, 00:13:22.173 "data_offset": 2048, 00:13:22.173 "data_size": 63488 00:13:22.173 }, 00:13:22.173 { 00:13:22.173 "name": "BaseBdev2", 00:13:22.173 "uuid": "5b5d431a-4f7d-49c5-b491-e91160973b8a", 00:13:22.173 "is_configured": true, 00:13:22.173 "data_offset": 2048, 00:13:22.173 "data_size": 63488 00:13:22.173 }, 00:13:22.173 { 00:13:22.173 "name": "BaseBdev3", 00:13:22.173 "uuid": "da5b5ffe-b96f-4adc-b06f-ce7f3284720f", 00:13:22.173 "is_configured": true, 00:13:22.173 "data_offset": 2048, 00:13:22.173 "data_size": 63488 00:13:22.173 } 00:13:22.173 ] 00:13:22.173 } 00:13:22.173 } 00:13:22.173 }' 00:13:22.173 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:22.434 BaseBdev2 00:13:22.434 BaseBdev3' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.434 14:12:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.434 [2024-09-30 14:12:27.038680] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.434 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.694 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.694 "name": "Existed_Raid", 00:13:22.694 "uuid": "2d914aa2-2c80-493a-8233-742c84d8df28", 00:13:22.694 "strip_size_kb": 64, 00:13:22.694 "state": "online", 00:13:22.694 "raid_level": "raid5f", 00:13:22.694 "superblock": true, 00:13:22.694 "num_base_bdevs": 3, 00:13:22.694 "num_base_bdevs_discovered": 2, 00:13:22.694 "num_base_bdevs_operational": 2, 00:13:22.694 "base_bdevs_list": [ 00:13:22.694 { 00:13:22.694 "name": null, 00:13:22.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:22.694 "is_configured": false, 00:13:22.694 "data_offset": 0, 00:13:22.694 "data_size": 63488 00:13:22.694 }, 00:13:22.694 { 00:13:22.694 "name": "BaseBdev2", 00:13:22.694 "uuid": "5b5d431a-4f7d-49c5-b491-e91160973b8a", 00:13:22.694 "is_configured": true, 00:13:22.694 "data_offset": 2048, 00:13:22.694 "data_size": 63488 00:13:22.694 }, 00:13:22.695 { 00:13:22.695 "name": "BaseBdev3", 00:13:22.695 "uuid": "da5b5ffe-b96f-4adc-b06f-ce7f3284720f", 00:13:22.695 "is_configured": true, 00:13:22.695 "data_offset": 2048, 00:13:22.695 "data_size": 63488 00:13:22.695 } 00:13:22.695 ] 00:13:22.695 }' 00:13:22.695 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.695 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.956 [2024-09-30 14:12:27.529024] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:22.956 [2024-09-30 14:12:27.529138] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:22.956 [2024-09-30 14:12:27.540054] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.956 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.956 [2024-09-30 14:12:27.599964] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:22.956 [2024-09-30 14:12:27.600041] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.216 BaseBdev2 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:23.216 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 [ 00:13:23.217 { 00:13:23.217 "name": "BaseBdev2", 00:13:23.217 "aliases": [ 00:13:23.217 "4fe2b167-ccf8-4192-8849-e21e6027ea1c" 00:13:23.217 ], 00:13:23.217 "product_name": "Malloc disk", 00:13:23.217 "block_size": 512, 00:13:23.217 "num_blocks": 65536, 00:13:23.217 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:23.217 "assigned_rate_limits": { 00:13:23.217 "rw_ios_per_sec": 0, 00:13:23.217 "rw_mbytes_per_sec": 0, 00:13:23.217 "r_mbytes_per_sec": 0, 00:13:23.217 "w_mbytes_per_sec": 0 00:13:23.217 }, 00:13:23.217 "claimed": false, 00:13:23.217 "zoned": false, 00:13:23.217 "supported_io_types": { 00:13:23.217 "read": true, 00:13:23.217 "write": true, 00:13:23.217 "unmap": true, 00:13:23.217 "flush": true, 00:13:23.217 "reset": true, 00:13:23.217 "nvme_admin": false, 00:13:23.217 "nvme_io": false, 00:13:23.217 "nvme_io_md": false, 00:13:23.217 "write_zeroes": true, 00:13:23.217 "zcopy": true, 00:13:23.217 "get_zone_info": false, 00:13:23.217 "zone_management": false, 00:13:23.217 "zone_append": false, 00:13:23.217 "compare": false, 00:13:23.217 "compare_and_write": false, 00:13:23.217 "abort": true, 00:13:23.217 "seek_hole": false, 00:13:23.217 "seek_data": false, 00:13:23.217 "copy": true, 00:13:23.217 "nvme_iov_md": false 00:13:23.217 }, 00:13:23.217 "memory_domains": [ 00:13:23.217 { 00:13:23.217 "dma_device_id": "system", 00:13:23.217 "dma_device_type": 1 00:13:23.217 }, 00:13:23.217 { 00:13:23.217 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:23.217 "dma_device_type": 2 00:13:23.217 } 00:13:23.217 ], 00:13:23.217 "driver_specific": {} 00:13:23.217 } 00:13:23.217 ] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 BaseBdev3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 [ 00:13:23.217 { 00:13:23.217 "name": "BaseBdev3", 00:13:23.217 "aliases": [ 00:13:23.217 "31f01d72-a5e0-4173-b9b1-dda318e49aec" 00:13:23.217 ], 00:13:23.217 "product_name": "Malloc disk", 00:13:23.217 "block_size": 512, 00:13:23.217 "num_blocks": 65536, 00:13:23.217 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:23.217 "assigned_rate_limits": { 00:13:23.217 "rw_ios_per_sec": 0, 00:13:23.217 "rw_mbytes_per_sec": 0, 00:13:23.217 "r_mbytes_per_sec": 0, 00:13:23.217 "w_mbytes_per_sec": 0 00:13:23.217 }, 00:13:23.217 "claimed": false, 00:13:23.217 "zoned": false, 00:13:23.217 "supported_io_types": { 00:13:23.217 "read": true, 00:13:23.217 "write": true, 00:13:23.217 "unmap": true, 00:13:23.217 "flush": true, 00:13:23.217 "reset": true, 00:13:23.217 "nvme_admin": false, 00:13:23.217 "nvme_io": false, 00:13:23.217 "nvme_io_md": false, 00:13:23.217 "write_zeroes": true, 00:13:23.217 "zcopy": true, 00:13:23.217 "get_zone_info": false, 00:13:23.217 "zone_management": false, 00:13:23.217 "zone_append": false, 00:13:23.217 "compare": false, 00:13:23.217 "compare_and_write": false, 00:13:23.217 "abort": true, 00:13:23.217 "seek_hole": false, 00:13:23.217 "seek_data": false, 00:13:23.217 "copy": true, 00:13:23.217 "nvme_iov_md": false 00:13:23.217 }, 00:13:23.217 "memory_domains": [ 00:13:23.217 { 00:13:23.217 "dma_device_id": "system", 00:13:23.217 "dma_device_type": 1 00:13:23.217 }, 00:13:23.217 { 00:13:23.217 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:23.217 "dma_device_type": 2 00:13:23.217 } 00:13:23.217 ], 00:13:23.217 "driver_specific": {} 00:13:23.217 } 00:13:23.217 ] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 [2024-09-30 14:12:27.773866] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:23.217 [2024-09-30 14:12:27.773949] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:23.217 [2024-09-30 14:12:27.773986] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:23.217 [2024-09-30 14:12:27.775708] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.217 "name": "Existed_Raid", 00:13:23.217 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:23.217 "strip_size_kb": 64, 00:13:23.217 "state": "configuring", 00:13:23.217 "raid_level": "raid5f", 00:13:23.217 "superblock": true, 00:13:23.217 "num_base_bdevs": 3, 00:13:23.217 "num_base_bdevs_discovered": 2, 00:13:23.217 "num_base_bdevs_operational": 3, 00:13:23.217 "base_bdevs_list": [ 00:13:23.217 { 00:13:23.217 "name": "BaseBdev1", 00:13:23.217 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.217 "is_configured": false, 00:13:23.217 "data_offset": 0, 00:13:23.217 "data_size": 0 00:13:23.217 }, 00:13:23.217 { 00:13:23.217 "name": "BaseBdev2", 00:13:23.217 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:23.217 "is_configured": true, 00:13:23.217 "data_offset": 2048, 00:13:23.217 "data_size": 63488 00:13:23.217 }, 00:13:23.217 { 00:13:23.217 "name": "BaseBdev3", 00:13:23.217 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:23.217 "is_configured": true, 00:13:23.217 "data_offset": 2048, 00:13:23.217 "data_size": 63488 00:13:23.217 } 00:13:23.217 ] 00:13:23.217 }' 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.217 14:12:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.800 [2024-09-30 14:12:28.229053] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.800 "name": "Existed_Raid", 00:13:23.800 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:23.800 "strip_size_kb": 64, 00:13:23.800 "state": "configuring", 00:13:23.800 "raid_level": "raid5f", 00:13:23.800 "superblock": true, 00:13:23.800 "num_base_bdevs": 3, 00:13:23.800 "num_base_bdevs_discovered": 1, 00:13:23.800 "num_base_bdevs_operational": 3, 00:13:23.800 "base_bdevs_list": [ 00:13:23.800 { 00:13:23.800 "name": "BaseBdev1", 00:13:23.800 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.800 "is_configured": false, 00:13:23.800 "data_offset": 0, 00:13:23.800 "data_size": 0 00:13:23.800 }, 00:13:23.800 { 00:13:23.800 "name": null, 00:13:23.800 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:23.800 "is_configured": false, 00:13:23.800 "data_offset": 0, 00:13:23.800 "data_size": 63488 00:13:23.800 }, 00:13:23.800 { 00:13:23.800 "name": "BaseBdev3", 00:13:23.800 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:23.800 "is_configured": true, 00:13:23.800 "data_offset": 2048, 00:13:23.800 "data_size": 63488 00:13:23.800 } 00:13:23.800 ] 00:13:23.800 }' 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.800 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.069 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.337 [2024-09-30 14:12:28.735112] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:24.337 BaseBdev1 00:13:24.337 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.337 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:24.337 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.338 [ 00:13:24.338 { 00:13:24.338 "name": "BaseBdev1", 00:13:24.338 "aliases": [ 00:13:24.338 "4054dd89-78db-4fc8-93a9-20e5b3751ac7" 00:13:24.338 ], 00:13:24.338 "product_name": "Malloc disk", 00:13:24.338 "block_size": 512, 00:13:24.338 "num_blocks": 65536, 00:13:24.338 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:24.338 "assigned_rate_limits": { 00:13:24.338 "rw_ios_per_sec": 0, 00:13:24.338 "rw_mbytes_per_sec": 0, 00:13:24.338 "r_mbytes_per_sec": 0, 00:13:24.338 "w_mbytes_per_sec": 0 00:13:24.338 }, 00:13:24.338 "claimed": true, 00:13:24.338 "claim_type": "exclusive_write", 00:13:24.338 "zoned": false, 00:13:24.338 "supported_io_types": { 00:13:24.338 "read": true, 00:13:24.338 "write": true, 00:13:24.338 "unmap": true, 00:13:24.338 "flush": true, 00:13:24.338 "reset": true, 00:13:24.338 "nvme_admin": false, 00:13:24.338 "nvme_io": false, 00:13:24.338 "nvme_io_md": false, 00:13:24.338 "write_zeroes": true, 00:13:24.338 "zcopy": true, 00:13:24.338 "get_zone_info": false, 00:13:24.338 "zone_management": false, 00:13:24.338 "zone_append": false, 00:13:24.338 "compare": false, 00:13:24.338 "compare_and_write": false, 00:13:24.338 "abort": true, 00:13:24.338 "seek_hole": false, 00:13:24.338 "seek_data": false, 00:13:24.338 "copy": true, 00:13:24.338 "nvme_iov_md": false 00:13:24.338 }, 00:13:24.338 "memory_domains": [ 00:13:24.338 { 00:13:24.338 "dma_device_id": "system", 00:13:24.338 "dma_device_type": 1 00:13:24.338 }, 00:13:24.338 { 00:13:24.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:24.338 "dma_device_type": 2 00:13:24.338 } 00:13:24.338 ], 00:13:24.338 "driver_specific": {} 00:13:24.338 } 00:13:24.338 ] 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.338 "name": "Existed_Raid", 00:13:24.338 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:24.338 "strip_size_kb": 64, 00:13:24.338 "state": "configuring", 00:13:24.338 "raid_level": "raid5f", 00:13:24.338 "superblock": true, 00:13:24.338 "num_base_bdevs": 3, 00:13:24.338 "num_base_bdevs_discovered": 2, 00:13:24.338 "num_base_bdevs_operational": 3, 00:13:24.338 "base_bdevs_list": [ 00:13:24.338 { 00:13:24.338 "name": "BaseBdev1", 00:13:24.338 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:24.338 "is_configured": true, 00:13:24.338 "data_offset": 2048, 00:13:24.338 "data_size": 63488 00:13:24.338 }, 00:13:24.338 { 00:13:24.338 "name": null, 00:13:24.338 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:24.338 "is_configured": false, 00:13:24.338 "data_offset": 0, 00:13:24.338 "data_size": 63488 00:13:24.338 }, 00:13:24.338 { 00:13:24.338 "name": "BaseBdev3", 00:13:24.338 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:24.338 "is_configured": true, 00:13:24.338 "data_offset": 2048, 00:13:24.338 "data_size": 63488 00:13:24.338 } 00:13:24.338 ] 00:13:24.338 }' 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.338 14:12:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.604 [2024-09-30 14:12:29.230293] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.604 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.863 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.863 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.863 "name": "Existed_Raid", 00:13:24.863 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:24.863 "strip_size_kb": 64, 00:13:24.863 "state": "configuring", 00:13:24.863 "raid_level": "raid5f", 00:13:24.863 "superblock": true, 00:13:24.863 "num_base_bdevs": 3, 00:13:24.863 "num_base_bdevs_discovered": 1, 00:13:24.863 "num_base_bdevs_operational": 3, 00:13:24.863 "base_bdevs_list": [ 00:13:24.863 { 00:13:24.863 "name": "BaseBdev1", 00:13:24.863 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:24.863 "is_configured": true, 00:13:24.863 "data_offset": 2048, 00:13:24.863 "data_size": 63488 00:13:24.863 }, 00:13:24.863 { 00:13:24.863 "name": null, 00:13:24.863 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:24.863 "is_configured": false, 00:13:24.863 "data_offset": 0, 00:13:24.863 "data_size": 63488 00:13:24.863 }, 00:13:24.863 { 00:13:24.863 "name": null, 00:13:24.863 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:24.863 "is_configured": false, 00:13:24.863 "data_offset": 0, 00:13:24.863 "data_size": 63488 00:13:24.863 } 00:13:24.863 ] 00:13:24.863 }' 00:13:24.864 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.864 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.122 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.122 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.122 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:25.122 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.122 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.123 [2024-09-30 14:12:29.733414] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.123 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.381 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.382 "name": "Existed_Raid", 00:13:25.382 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:25.382 "strip_size_kb": 64, 00:13:25.382 "state": "configuring", 00:13:25.382 "raid_level": "raid5f", 00:13:25.382 "superblock": true, 00:13:25.382 "num_base_bdevs": 3, 00:13:25.382 "num_base_bdevs_discovered": 2, 00:13:25.382 "num_base_bdevs_operational": 3, 00:13:25.382 "base_bdevs_list": [ 00:13:25.382 { 00:13:25.382 "name": "BaseBdev1", 00:13:25.382 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:25.382 "is_configured": true, 00:13:25.382 "data_offset": 2048, 00:13:25.382 "data_size": 63488 00:13:25.382 }, 00:13:25.382 { 00:13:25.382 "name": null, 00:13:25.382 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:25.382 "is_configured": false, 00:13:25.382 "data_offset": 0, 00:13:25.382 "data_size": 63488 00:13:25.382 }, 00:13:25.382 { 00:13:25.382 "name": "BaseBdev3", 00:13:25.382 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:25.382 "is_configured": true, 00:13:25.382 "data_offset": 2048, 00:13:25.382 "data_size": 63488 00:13:25.382 } 00:13:25.382 ] 00:13:25.382 }' 00:13:25.382 14:12:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.382 14:12:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.703 [2024-09-30 14:12:30.212611] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.703 "name": "Existed_Raid", 00:13:25.703 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:25.703 "strip_size_kb": 64, 00:13:25.703 "state": "configuring", 00:13:25.703 "raid_level": "raid5f", 00:13:25.703 "superblock": true, 00:13:25.703 "num_base_bdevs": 3, 00:13:25.703 "num_base_bdevs_discovered": 1, 00:13:25.703 "num_base_bdevs_operational": 3, 00:13:25.703 "base_bdevs_list": [ 00:13:25.703 { 00:13:25.703 "name": null, 00:13:25.703 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:25.703 "is_configured": false, 00:13:25.703 "data_offset": 0, 00:13:25.703 "data_size": 63488 00:13:25.703 }, 00:13:25.703 { 00:13:25.703 "name": null, 00:13:25.703 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:25.703 "is_configured": false, 00:13:25.703 "data_offset": 0, 00:13:25.703 "data_size": 63488 00:13:25.703 }, 00:13:25.703 { 00:13:25.703 "name": "BaseBdev3", 00:13:25.703 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:25.703 "is_configured": true, 00:13:25.703 "data_offset": 2048, 00:13:25.703 "data_size": 63488 00:13:25.703 } 00:13:25.703 ] 00:13:25.703 }' 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.703 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.270 [2024-09-30 14:12:30.702204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.270 "name": "Existed_Raid", 00:13:26.270 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:26.270 "strip_size_kb": 64, 00:13:26.270 "state": "configuring", 00:13:26.270 "raid_level": "raid5f", 00:13:26.270 "superblock": true, 00:13:26.270 "num_base_bdevs": 3, 00:13:26.270 "num_base_bdevs_discovered": 2, 00:13:26.270 "num_base_bdevs_operational": 3, 00:13:26.270 "base_bdevs_list": [ 00:13:26.270 { 00:13:26.270 "name": null, 00:13:26.270 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:26.270 "is_configured": false, 00:13:26.270 "data_offset": 0, 00:13:26.270 "data_size": 63488 00:13:26.270 }, 00:13:26.270 { 00:13:26.270 "name": "BaseBdev2", 00:13:26.270 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:26.270 "is_configured": true, 00:13:26.270 "data_offset": 2048, 00:13:26.270 "data_size": 63488 00:13:26.270 }, 00:13:26.270 { 00:13:26.270 "name": "BaseBdev3", 00:13:26.270 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:26.270 "is_configured": true, 00:13:26.270 "data_offset": 2048, 00:13:26.270 "data_size": 63488 00:13:26.270 } 00:13:26.270 ] 00:13:26.270 }' 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.270 14:12:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.528 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.528 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:26.528 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.528 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 4054dd89-78db-4fc8-93a9-20e5b3751ac7 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.787 [2024-09-30 14:12:31.283942] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:26.787 [2024-09-30 14:12:31.284098] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:13:26.787 [2024-09-30 14:12:31.284111] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:26.787 [2024-09-30 14:12:31.284354] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:13:26.787 NewBaseBdev 00:13:26.787 [2024-09-30 14:12:31.284760] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:13:26.787 [2024-09-30 14:12:31.284783] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:13:26.787 [2024-09-30 14:12:31.284883] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.787 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.787 [ 00:13:26.787 { 00:13:26.787 "name": "NewBaseBdev", 00:13:26.787 "aliases": [ 00:13:26.787 "4054dd89-78db-4fc8-93a9-20e5b3751ac7" 00:13:26.787 ], 00:13:26.787 "product_name": "Malloc disk", 00:13:26.787 "block_size": 512, 00:13:26.787 "num_blocks": 65536, 00:13:26.787 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:26.787 "assigned_rate_limits": { 00:13:26.787 "rw_ios_per_sec": 0, 00:13:26.787 "rw_mbytes_per_sec": 0, 00:13:26.787 "r_mbytes_per_sec": 0, 00:13:26.788 "w_mbytes_per_sec": 0 00:13:26.788 }, 00:13:26.788 "claimed": true, 00:13:26.788 "claim_type": "exclusive_write", 00:13:26.788 "zoned": false, 00:13:26.788 "supported_io_types": { 00:13:26.788 "read": true, 00:13:26.788 "write": true, 00:13:26.788 "unmap": true, 00:13:26.788 "flush": true, 00:13:26.788 "reset": true, 00:13:26.788 "nvme_admin": false, 00:13:26.788 "nvme_io": false, 00:13:26.788 "nvme_io_md": false, 00:13:26.788 "write_zeroes": true, 00:13:26.788 "zcopy": true, 00:13:26.788 "get_zone_info": false, 00:13:26.788 "zone_management": false, 00:13:26.788 "zone_append": false, 00:13:26.788 "compare": false, 00:13:26.788 "compare_and_write": false, 00:13:26.788 "abort": true, 00:13:26.788 "seek_hole": false, 00:13:26.788 "seek_data": false, 00:13:26.788 "copy": true, 00:13:26.788 "nvme_iov_md": false 00:13:26.788 }, 00:13:26.788 "memory_domains": [ 00:13:26.788 { 00:13:26.788 "dma_device_id": "system", 00:13:26.788 "dma_device_type": 1 00:13:26.788 }, 00:13:26.788 { 00:13:26.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:26.788 "dma_device_type": 2 00:13:26.788 } 00:13:26.788 ], 00:13:26.788 "driver_specific": {} 00:13:26.788 } 00:13:26.788 ] 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.788 "name": "Existed_Raid", 00:13:26.788 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:26.788 "strip_size_kb": 64, 00:13:26.788 "state": "online", 00:13:26.788 "raid_level": "raid5f", 00:13:26.788 "superblock": true, 00:13:26.788 "num_base_bdevs": 3, 00:13:26.788 "num_base_bdevs_discovered": 3, 00:13:26.788 "num_base_bdevs_operational": 3, 00:13:26.788 "base_bdevs_list": [ 00:13:26.788 { 00:13:26.788 "name": "NewBaseBdev", 00:13:26.788 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:26.788 "is_configured": true, 00:13:26.788 "data_offset": 2048, 00:13:26.788 "data_size": 63488 00:13:26.788 }, 00:13:26.788 { 00:13:26.788 "name": "BaseBdev2", 00:13:26.788 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:26.788 "is_configured": true, 00:13:26.788 "data_offset": 2048, 00:13:26.788 "data_size": 63488 00:13:26.788 }, 00:13:26.788 { 00:13:26.788 "name": "BaseBdev3", 00:13:26.788 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:26.788 "is_configured": true, 00:13:26.788 "data_offset": 2048, 00:13:26.788 "data_size": 63488 00:13:26.788 } 00:13:26.788 ] 00:13:26.788 }' 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.788 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.356 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.357 [2024-09-30 14:12:31.771435] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:27.357 "name": "Existed_Raid", 00:13:27.357 "aliases": [ 00:13:27.357 "2df9d33d-617c-4b24-a063-4cca5f2c5a8f" 00:13:27.357 ], 00:13:27.357 "product_name": "Raid Volume", 00:13:27.357 "block_size": 512, 00:13:27.357 "num_blocks": 126976, 00:13:27.357 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:27.357 "assigned_rate_limits": { 00:13:27.357 "rw_ios_per_sec": 0, 00:13:27.357 "rw_mbytes_per_sec": 0, 00:13:27.357 "r_mbytes_per_sec": 0, 00:13:27.357 "w_mbytes_per_sec": 0 00:13:27.357 }, 00:13:27.357 "claimed": false, 00:13:27.357 "zoned": false, 00:13:27.357 "supported_io_types": { 00:13:27.357 "read": true, 00:13:27.357 "write": true, 00:13:27.357 "unmap": false, 00:13:27.357 "flush": false, 00:13:27.357 "reset": true, 00:13:27.357 "nvme_admin": false, 00:13:27.357 "nvme_io": false, 00:13:27.357 "nvme_io_md": false, 00:13:27.357 "write_zeroes": true, 00:13:27.357 "zcopy": false, 00:13:27.357 "get_zone_info": false, 00:13:27.357 "zone_management": false, 00:13:27.357 "zone_append": false, 00:13:27.357 "compare": false, 00:13:27.357 "compare_and_write": false, 00:13:27.357 "abort": false, 00:13:27.357 "seek_hole": false, 00:13:27.357 "seek_data": false, 00:13:27.357 "copy": false, 00:13:27.357 "nvme_iov_md": false 00:13:27.357 }, 00:13:27.357 "driver_specific": { 00:13:27.357 "raid": { 00:13:27.357 "uuid": "2df9d33d-617c-4b24-a063-4cca5f2c5a8f", 00:13:27.357 "strip_size_kb": 64, 00:13:27.357 "state": "online", 00:13:27.357 "raid_level": "raid5f", 00:13:27.357 "superblock": true, 00:13:27.357 "num_base_bdevs": 3, 00:13:27.357 "num_base_bdevs_discovered": 3, 00:13:27.357 "num_base_bdevs_operational": 3, 00:13:27.357 "base_bdevs_list": [ 00:13:27.357 { 00:13:27.357 "name": "NewBaseBdev", 00:13:27.357 "uuid": "4054dd89-78db-4fc8-93a9-20e5b3751ac7", 00:13:27.357 "is_configured": true, 00:13:27.357 "data_offset": 2048, 00:13:27.357 "data_size": 63488 00:13:27.357 }, 00:13:27.357 { 00:13:27.357 "name": "BaseBdev2", 00:13:27.357 "uuid": "4fe2b167-ccf8-4192-8849-e21e6027ea1c", 00:13:27.357 "is_configured": true, 00:13:27.357 "data_offset": 2048, 00:13:27.357 "data_size": 63488 00:13:27.357 }, 00:13:27.357 { 00:13:27.357 "name": "BaseBdev3", 00:13:27.357 "uuid": "31f01d72-a5e0-4173-b9b1-dda318e49aec", 00:13:27.357 "is_configured": true, 00:13:27.357 "data_offset": 2048, 00:13:27.357 "data_size": 63488 00:13:27.357 } 00:13:27.357 ] 00:13:27.357 } 00:13:27.357 } 00:13:27.357 }' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:27.357 BaseBdev2 00:13:27.357 BaseBdev3' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.357 14:12:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.357 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:27.357 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:27.357 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.617 [2024-09-30 14:12:32.050800] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:27.617 [2024-09-30 14:12:32.050863] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:27.617 [2024-09-30 14:12:32.050926] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:27.617 [2024-09-30 14:12:32.051177] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:27.617 [2024-09-30 14:12:32.051187] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 92079 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 92079 ']' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 92079 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 92079 00:13:27.617 killing process with pid 92079 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 92079' 00:13:27.617 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 92079 00:13:27.618 [2024-09-30 14:12:32.085676] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:27.618 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 92079 00:13:27.618 [2024-09-30 14:12:32.116398] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:27.878 14:12:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:27.878 00:13:27.878 real 0m9.050s 00:13:27.878 user 0m15.423s 00:13:27.878 sys 0m1.950s 00:13:27.878 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:27.878 14:12:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:27.878 ************************************ 00:13:27.878 END TEST raid5f_state_function_test_sb 00:13:27.878 ************************************ 00:13:27.878 14:12:32 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:13:27.878 14:12:32 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:13:27.878 14:12:32 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:27.878 14:12:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:27.878 ************************************ 00:13:27.878 START TEST raid5f_superblock_test 00:13:27.878 ************************************ 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 3 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=92677 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 92677 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 92677 ']' 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:27.878 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:27.878 14:12:32 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.138 [2024-09-30 14:12:32.549335] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:13:28.138 [2024-09-30 14:12:32.549578] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92677 ] 00:13:28.138 [2024-09-30 14:12:32.687447] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:13:28.138 [2024-09-30 14:12:32.716808] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:28.138 [2024-09-30 14:12:32.760572] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:28.397 [2024-09-30 14:12:32.802740] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:28.397 [2024-09-30 14:12:32.802777] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 malloc1 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 [2024-09-30 14:12:33.388759] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:28.967 [2024-09-30 14:12:33.388859] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:28.967 [2024-09-30 14:12:33.388904] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:28.967 [2024-09-30 14:12:33.388946] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:28.967 [2024-09-30 14:12:33.390908] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:28.967 [2024-09-30 14:12:33.390974] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:28.967 pt1 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 malloc2 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 [2024-09-30 14:12:33.430124] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:28.967 [2024-09-30 14:12:33.430234] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:28.967 [2024-09-30 14:12:33.430261] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:28.967 [2024-09-30 14:12:33.430272] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:28.967 [2024-09-30 14:12:33.432280] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:28.967 [2024-09-30 14:12:33.432315] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:28.967 pt2 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 malloc3 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 [2024-09-30 14:12:33.458436] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:28.967 [2024-09-30 14:12:33.458541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:28.967 [2024-09-30 14:12:33.458577] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:28.967 [2024-09-30 14:12:33.458600] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:28.967 [2024-09-30 14:12:33.460609] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:28.967 [2024-09-30 14:12:33.460675] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:28.967 pt3 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.967 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.967 [2024-09-30 14:12:33.470473] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:28.967 [2024-09-30 14:12:33.472232] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:28.968 [2024-09-30 14:12:33.472325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:28.968 [2024-09-30 14:12:33.472491] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:28.968 [2024-09-30 14:12:33.472550] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:28.968 [2024-09-30 14:12:33.472797] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:13:28.968 [2024-09-30 14:12:33.473198] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:28.968 [2024-09-30 14:12:33.473240] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:28.968 [2024-09-30 14:12:33.473398] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.968 "name": "raid_bdev1", 00:13:28.968 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:28.968 "strip_size_kb": 64, 00:13:28.968 "state": "online", 00:13:28.968 "raid_level": "raid5f", 00:13:28.968 "superblock": true, 00:13:28.968 "num_base_bdevs": 3, 00:13:28.968 "num_base_bdevs_discovered": 3, 00:13:28.968 "num_base_bdevs_operational": 3, 00:13:28.968 "base_bdevs_list": [ 00:13:28.968 { 00:13:28.968 "name": "pt1", 00:13:28.968 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:28.968 "is_configured": true, 00:13:28.968 "data_offset": 2048, 00:13:28.968 "data_size": 63488 00:13:28.968 }, 00:13:28.968 { 00:13:28.968 "name": "pt2", 00:13:28.968 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:28.968 "is_configured": true, 00:13:28.968 "data_offset": 2048, 00:13:28.968 "data_size": 63488 00:13:28.968 }, 00:13:28.968 { 00:13:28.968 "name": "pt3", 00:13:28.968 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:28.968 "is_configured": true, 00:13:28.968 "data_offset": 2048, 00:13:28.968 "data_size": 63488 00:13:28.968 } 00:13:28.968 ] 00:13:28.968 }' 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.968 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.537 [2024-09-30 14:12:33.950081] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:29.537 "name": "raid_bdev1", 00:13:29.537 "aliases": [ 00:13:29.537 "95b1da33-3157-48b7-88ae-d6faf9e4d59a" 00:13:29.537 ], 00:13:29.537 "product_name": "Raid Volume", 00:13:29.537 "block_size": 512, 00:13:29.537 "num_blocks": 126976, 00:13:29.537 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:29.537 "assigned_rate_limits": { 00:13:29.537 "rw_ios_per_sec": 0, 00:13:29.537 "rw_mbytes_per_sec": 0, 00:13:29.537 "r_mbytes_per_sec": 0, 00:13:29.537 "w_mbytes_per_sec": 0 00:13:29.537 }, 00:13:29.537 "claimed": false, 00:13:29.537 "zoned": false, 00:13:29.537 "supported_io_types": { 00:13:29.537 "read": true, 00:13:29.537 "write": true, 00:13:29.537 "unmap": false, 00:13:29.537 "flush": false, 00:13:29.537 "reset": true, 00:13:29.537 "nvme_admin": false, 00:13:29.537 "nvme_io": false, 00:13:29.537 "nvme_io_md": false, 00:13:29.537 "write_zeroes": true, 00:13:29.537 "zcopy": false, 00:13:29.537 "get_zone_info": false, 00:13:29.537 "zone_management": false, 00:13:29.537 "zone_append": false, 00:13:29.537 "compare": false, 00:13:29.537 "compare_and_write": false, 00:13:29.537 "abort": false, 00:13:29.537 "seek_hole": false, 00:13:29.537 "seek_data": false, 00:13:29.537 "copy": false, 00:13:29.537 "nvme_iov_md": false 00:13:29.537 }, 00:13:29.537 "driver_specific": { 00:13:29.537 "raid": { 00:13:29.537 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:29.537 "strip_size_kb": 64, 00:13:29.537 "state": "online", 00:13:29.537 "raid_level": "raid5f", 00:13:29.537 "superblock": true, 00:13:29.537 "num_base_bdevs": 3, 00:13:29.537 "num_base_bdevs_discovered": 3, 00:13:29.537 "num_base_bdevs_operational": 3, 00:13:29.537 "base_bdevs_list": [ 00:13:29.537 { 00:13:29.537 "name": "pt1", 00:13:29.537 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:29.537 "is_configured": true, 00:13:29.537 "data_offset": 2048, 00:13:29.537 "data_size": 63488 00:13:29.537 }, 00:13:29.537 { 00:13:29.537 "name": "pt2", 00:13:29.537 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:29.537 "is_configured": true, 00:13:29.537 "data_offset": 2048, 00:13:29.537 "data_size": 63488 00:13:29.537 }, 00:13:29.537 { 00:13:29.537 "name": "pt3", 00:13:29.537 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:29.537 "is_configured": true, 00:13:29.537 "data_offset": 2048, 00:13:29.537 "data_size": 63488 00:13:29.537 } 00:13:29.537 ] 00:13:29.537 } 00:13:29.537 } 00:13:29.537 }' 00:13:29.537 14:12:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:29.537 pt2 00:13:29.537 pt3' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.537 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 [2024-09-30 14:12:34.253521] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=95b1da33-3157-48b7-88ae-d6faf9e4d59a 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 95b1da33-3157-48b7-88ae-d6faf9e4d59a ']' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 [2024-09-30 14:12:34.297285] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:29.798 [2024-09-30 14:12:34.297350] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:29.798 [2024-09-30 14:12:34.297434] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:29.798 [2024-09-30 14:12:34.297525] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:29.798 [2024-09-30 14:12:34.297559] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.798 [2024-09-30 14:12:34.441073] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:29.798 [2024-09-30 14:12:34.442846] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:29.798 [2024-09-30 14:12:34.442895] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:29.798 [2024-09-30 14:12:34.442933] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:29.798 [2024-09-30 14:12:34.442986] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:29.798 [2024-09-30 14:12:34.443004] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:29.798 [2024-09-30 14:12:34.443028] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:29.798 [2024-09-30 14:12:34.443036] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:13:29.798 request: 00:13:29.798 { 00:13:29.798 "name": "raid_bdev1", 00:13:29.798 "raid_level": "raid5f", 00:13:29.798 "base_bdevs": [ 00:13:29.798 "malloc1", 00:13:29.798 "malloc2", 00:13:29.798 "malloc3" 00:13:29.798 ], 00:13:29.798 "strip_size_kb": 64, 00:13:29.798 "superblock": false, 00:13:29.798 "method": "bdev_raid_create", 00:13:29.798 "req_id": 1 00:13:29.798 } 00:13:29.798 Got JSON-RPC error response 00:13:29.798 response: 00:13:29.798 { 00:13:29.798 "code": -17, 00:13:29.798 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:29.798 } 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:29.798 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.058 [2024-09-30 14:12:34.488957] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:30.058 [2024-09-30 14:12:34.489043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.058 [2024-09-30 14:12:34.489075] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:30.058 [2024-09-30 14:12:34.489099] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.058 [2024-09-30 14:12:34.491017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.058 [2024-09-30 14:12:34.491080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:30.058 [2024-09-30 14:12:34.491155] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:30.058 [2024-09-30 14:12:34.491221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:30.058 pt1 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.058 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.059 "name": "raid_bdev1", 00:13:30.059 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:30.059 "strip_size_kb": 64, 00:13:30.059 "state": "configuring", 00:13:30.059 "raid_level": "raid5f", 00:13:30.059 "superblock": true, 00:13:30.059 "num_base_bdevs": 3, 00:13:30.059 "num_base_bdevs_discovered": 1, 00:13:30.059 "num_base_bdevs_operational": 3, 00:13:30.059 "base_bdevs_list": [ 00:13:30.059 { 00:13:30.059 "name": "pt1", 00:13:30.059 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:30.059 "is_configured": true, 00:13:30.059 "data_offset": 2048, 00:13:30.059 "data_size": 63488 00:13:30.059 }, 00:13:30.059 { 00:13:30.059 "name": null, 00:13:30.059 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:30.059 "is_configured": false, 00:13:30.059 "data_offset": 2048, 00:13:30.059 "data_size": 63488 00:13:30.059 }, 00:13:30.059 { 00:13:30.059 "name": null, 00:13:30.059 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:30.059 "is_configured": false, 00:13:30.059 "data_offset": 2048, 00:13:30.059 "data_size": 63488 00:13:30.059 } 00:13:30.059 ] 00:13:30.059 }' 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.059 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.319 [2024-09-30 14:12:34.908233] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:30.319 [2024-09-30 14:12:34.908293] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.319 [2024-09-30 14:12:34.908312] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:30.319 [2024-09-30 14:12:34.908320] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.319 [2024-09-30 14:12:34.908626] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.319 [2024-09-30 14:12:34.908648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:30.319 [2024-09-30 14:12:34.908700] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:30.319 [2024-09-30 14:12:34.908716] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:30.319 pt2 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.319 [2024-09-30 14:12:34.920244] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.319 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.578 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.578 "name": "raid_bdev1", 00:13:30.578 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:30.578 "strip_size_kb": 64, 00:13:30.578 "state": "configuring", 00:13:30.578 "raid_level": "raid5f", 00:13:30.578 "superblock": true, 00:13:30.578 "num_base_bdevs": 3, 00:13:30.578 "num_base_bdevs_discovered": 1, 00:13:30.578 "num_base_bdevs_operational": 3, 00:13:30.578 "base_bdevs_list": [ 00:13:30.578 { 00:13:30.578 "name": "pt1", 00:13:30.578 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:30.578 "is_configured": true, 00:13:30.578 "data_offset": 2048, 00:13:30.578 "data_size": 63488 00:13:30.578 }, 00:13:30.578 { 00:13:30.579 "name": null, 00:13:30.579 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:30.579 "is_configured": false, 00:13:30.579 "data_offset": 0, 00:13:30.579 "data_size": 63488 00:13:30.579 }, 00:13:30.579 { 00:13:30.579 "name": null, 00:13:30.579 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:30.579 "is_configured": false, 00:13:30.579 "data_offset": 2048, 00:13:30.579 "data_size": 63488 00:13:30.579 } 00:13:30.579 ] 00:13:30.579 }' 00:13:30.579 14:12:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.579 14:12:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.839 [2024-09-30 14:12:35.391365] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:30.839 [2024-09-30 14:12:35.391416] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.839 [2024-09-30 14:12:35.391428] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:30.839 [2024-09-30 14:12:35.391437] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.839 [2024-09-30 14:12:35.391738] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.839 [2024-09-30 14:12:35.391758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:30.839 [2024-09-30 14:12:35.391806] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:30.839 [2024-09-30 14:12:35.391832] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:30.839 pt2 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.839 [2024-09-30 14:12:35.403358] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:30.839 [2024-09-30 14:12:35.403404] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.839 [2024-09-30 14:12:35.403416] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:30.839 [2024-09-30 14:12:35.403425] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.839 [2024-09-30 14:12:35.403730] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.839 [2024-09-30 14:12:35.403749] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:30.839 [2024-09-30 14:12:35.403795] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:30.839 [2024-09-30 14:12:35.403819] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:30.839 [2024-09-30 14:12:35.403900] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:13:30.839 [2024-09-30 14:12:35.403918] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:30.839 [2024-09-30 14:12:35.404115] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:13:30.839 [2024-09-30 14:12:35.404520] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:13:30.839 [2024-09-30 14:12:35.404538] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:13:30.839 [2024-09-30 14:12:35.404629] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:30.839 pt3 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.839 "name": "raid_bdev1", 00:13:30.839 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:30.839 "strip_size_kb": 64, 00:13:30.839 "state": "online", 00:13:30.839 "raid_level": "raid5f", 00:13:30.839 "superblock": true, 00:13:30.839 "num_base_bdevs": 3, 00:13:30.839 "num_base_bdevs_discovered": 3, 00:13:30.839 "num_base_bdevs_operational": 3, 00:13:30.839 "base_bdevs_list": [ 00:13:30.839 { 00:13:30.839 "name": "pt1", 00:13:30.839 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:30.839 "is_configured": true, 00:13:30.839 "data_offset": 2048, 00:13:30.839 "data_size": 63488 00:13:30.839 }, 00:13:30.839 { 00:13:30.839 "name": "pt2", 00:13:30.839 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:30.839 "is_configured": true, 00:13:30.839 "data_offset": 2048, 00:13:30.839 "data_size": 63488 00:13:30.839 }, 00:13:30.839 { 00:13:30.839 "name": "pt3", 00:13:30.839 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:30.839 "is_configured": true, 00:13:30.839 "data_offset": 2048, 00:13:30.839 "data_size": 63488 00:13:30.839 } 00:13:30.839 ] 00:13:30.839 }' 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.839 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.407 [2024-09-30 14:12:35.875052] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:31.407 "name": "raid_bdev1", 00:13:31.407 "aliases": [ 00:13:31.407 "95b1da33-3157-48b7-88ae-d6faf9e4d59a" 00:13:31.407 ], 00:13:31.407 "product_name": "Raid Volume", 00:13:31.407 "block_size": 512, 00:13:31.407 "num_blocks": 126976, 00:13:31.407 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:31.407 "assigned_rate_limits": { 00:13:31.407 "rw_ios_per_sec": 0, 00:13:31.407 "rw_mbytes_per_sec": 0, 00:13:31.407 "r_mbytes_per_sec": 0, 00:13:31.407 "w_mbytes_per_sec": 0 00:13:31.407 }, 00:13:31.407 "claimed": false, 00:13:31.407 "zoned": false, 00:13:31.407 "supported_io_types": { 00:13:31.407 "read": true, 00:13:31.407 "write": true, 00:13:31.407 "unmap": false, 00:13:31.407 "flush": false, 00:13:31.407 "reset": true, 00:13:31.407 "nvme_admin": false, 00:13:31.407 "nvme_io": false, 00:13:31.407 "nvme_io_md": false, 00:13:31.407 "write_zeroes": true, 00:13:31.407 "zcopy": false, 00:13:31.407 "get_zone_info": false, 00:13:31.407 "zone_management": false, 00:13:31.407 "zone_append": false, 00:13:31.407 "compare": false, 00:13:31.407 "compare_and_write": false, 00:13:31.407 "abort": false, 00:13:31.407 "seek_hole": false, 00:13:31.407 "seek_data": false, 00:13:31.407 "copy": false, 00:13:31.407 "nvme_iov_md": false 00:13:31.407 }, 00:13:31.407 "driver_specific": { 00:13:31.407 "raid": { 00:13:31.407 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:31.407 "strip_size_kb": 64, 00:13:31.407 "state": "online", 00:13:31.407 "raid_level": "raid5f", 00:13:31.407 "superblock": true, 00:13:31.407 "num_base_bdevs": 3, 00:13:31.407 "num_base_bdevs_discovered": 3, 00:13:31.407 "num_base_bdevs_operational": 3, 00:13:31.407 "base_bdevs_list": [ 00:13:31.407 { 00:13:31.407 "name": "pt1", 00:13:31.407 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:31.407 "is_configured": true, 00:13:31.407 "data_offset": 2048, 00:13:31.407 "data_size": 63488 00:13:31.407 }, 00:13:31.407 { 00:13:31.407 "name": "pt2", 00:13:31.407 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:31.407 "is_configured": true, 00:13:31.407 "data_offset": 2048, 00:13:31.407 "data_size": 63488 00:13:31.407 }, 00:13:31.407 { 00:13:31.407 "name": "pt3", 00:13:31.407 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:31.407 "is_configured": true, 00:13:31.407 "data_offset": 2048, 00:13:31.407 "data_size": 63488 00:13:31.407 } 00:13:31.407 ] 00:13:31.407 } 00:13:31.407 } 00:13:31.407 }' 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:31.407 pt2 00:13:31.407 pt3' 00:13:31.407 14:12:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:31.407 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:31.408 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:31.408 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:31.408 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.408 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.408 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.666 [2024-09-30 14:12:36.142674] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 95b1da33-3157-48b7-88ae-d6faf9e4d59a '!=' 95b1da33-3157-48b7-88ae-d6faf9e4d59a ']' 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.666 [2024-09-30 14:12:36.190456] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:31.666 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:31.667 "name": "raid_bdev1", 00:13:31.667 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:31.667 "strip_size_kb": 64, 00:13:31.667 "state": "online", 00:13:31.667 "raid_level": "raid5f", 00:13:31.667 "superblock": true, 00:13:31.667 "num_base_bdevs": 3, 00:13:31.667 "num_base_bdevs_discovered": 2, 00:13:31.667 "num_base_bdevs_operational": 2, 00:13:31.667 "base_bdevs_list": [ 00:13:31.667 { 00:13:31.667 "name": null, 00:13:31.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.667 "is_configured": false, 00:13:31.667 "data_offset": 0, 00:13:31.667 "data_size": 63488 00:13:31.667 }, 00:13:31.667 { 00:13:31.667 "name": "pt2", 00:13:31.667 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:31.667 "is_configured": true, 00:13:31.667 "data_offset": 2048, 00:13:31.667 "data_size": 63488 00:13:31.667 }, 00:13:31.667 { 00:13:31.667 "name": "pt3", 00:13:31.667 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:31.667 "is_configured": true, 00:13:31.667 "data_offset": 2048, 00:13:31.667 "data_size": 63488 00:13:31.667 } 00:13:31.667 ] 00:13:31.667 }' 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:31.667 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.927 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:31.927 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.927 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.927 [2024-09-30 14:12:36.573752] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:31.927 [2024-09-30 14:12:36.573817] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:31.927 [2024-09-30 14:12:36.573880] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:31.927 [2024-09-30 14:12:36.573935] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:31.927 [2024-09-30 14:12:36.573967] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:13:31.927 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:32.187 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.188 [2024-09-30 14:12:36.661589] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:32.188 [2024-09-30 14:12:36.661677] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.188 [2024-09-30 14:12:36.661694] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:13:32.188 [2024-09-30 14:12:36.661704] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.188 [2024-09-30 14:12:36.663778] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.188 [2024-09-30 14:12:36.663815] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:32.188 [2024-09-30 14:12:36.663867] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:32.188 [2024-09-30 14:12:36.663898] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:32.188 pt2 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.188 "name": "raid_bdev1", 00:13:32.188 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:32.188 "strip_size_kb": 64, 00:13:32.188 "state": "configuring", 00:13:32.188 "raid_level": "raid5f", 00:13:32.188 "superblock": true, 00:13:32.188 "num_base_bdevs": 3, 00:13:32.188 "num_base_bdevs_discovered": 1, 00:13:32.188 "num_base_bdevs_operational": 2, 00:13:32.188 "base_bdevs_list": [ 00:13:32.188 { 00:13:32.188 "name": null, 00:13:32.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.188 "is_configured": false, 00:13:32.188 "data_offset": 2048, 00:13:32.188 "data_size": 63488 00:13:32.188 }, 00:13:32.188 { 00:13:32.188 "name": "pt2", 00:13:32.188 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:32.188 "is_configured": true, 00:13:32.188 "data_offset": 2048, 00:13:32.188 "data_size": 63488 00:13:32.188 }, 00:13:32.188 { 00:13:32.188 "name": null, 00:13:32.188 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:32.188 "is_configured": false, 00:13:32.188 "data_offset": 2048, 00:13:32.188 "data_size": 63488 00:13:32.188 } 00:13:32.188 ] 00:13:32.188 }' 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.188 14:12:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.448 [2024-09-30 14:12:37.088881] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:32.448 [2024-09-30 14:12:37.088972] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.448 [2024-09-30 14:12:37.089003] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:13:32.448 [2024-09-30 14:12:37.089033] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.448 [2024-09-30 14:12:37.089343] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.448 [2024-09-30 14:12:37.089401] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:32.448 [2024-09-30 14:12:37.089478] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:32.448 [2024-09-30 14:12:37.089543] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:32.448 [2024-09-30 14:12:37.089647] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:13:32.448 [2024-09-30 14:12:37.089686] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:32.448 [2024-09-30 14:12:37.089910] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:32.448 [2024-09-30 14:12:37.090350] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:13:32.448 [2024-09-30 14:12:37.090395] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:13:32.448 [2024-09-30 14:12:37.090689] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:32.448 pt3 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.448 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.708 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.708 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.708 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.708 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.708 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.708 "name": "raid_bdev1", 00:13:32.708 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:32.708 "strip_size_kb": 64, 00:13:32.708 "state": "online", 00:13:32.708 "raid_level": "raid5f", 00:13:32.708 "superblock": true, 00:13:32.708 "num_base_bdevs": 3, 00:13:32.708 "num_base_bdevs_discovered": 2, 00:13:32.708 "num_base_bdevs_operational": 2, 00:13:32.708 "base_bdevs_list": [ 00:13:32.708 { 00:13:32.708 "name": null, 00:13:32.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.708 "is_configured": false, 00:13:32.708 "data_offset": 2048, 00:13:32.708 "data_size": 63488 00:13:32.708 }, 00:13:32.708 { 00:13:32.708 "name": "pt2", 00:13:32.708 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:32.708 "is_configured": true, 00:13:32.708 "data_offset": 2048, 00:13:32.708 "data_size": 63488 00:13:32.708 }, 00:13:32.708 { 00:13:32.708 "name": "pt3", 00:13:32.709 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:32.709 "is_configured": true, 00:13:32.709 "data_offset": 2048, 00:13:32.709 "data_size": 63488 00:13:32.709 } 00:13:32.709 ] 00:13:32.709 }' 00:13:32.709 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.709 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 [2024-09-30 14:12:37.552107] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:32.968 [2024-09-30 14:12:37.552178] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:32.968 [2024-09-30 14:12:37.552240] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:32.968 [2024-09-30 14:12:37.552298] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:32.968 [2024-09-30 14:12:37.552328] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:32.968 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.969 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.228 [2024-09-30 14:12:37.628000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:33.228 [2024-09-30 14:12:37.628048] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:33.228 [2024-09-30 14:12:37.628065] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:33.228 [2024-09-30 14:12:37.628073] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:33.228 [2024-09-30 14:12:37.630146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:33.228 [2024-09-30 14:12:37.630214] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:33.228 [2024-09-30 14:12:37.630274] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:33.228 [2024-09-30 14:12:37.630301] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:33.228 [2024-09-30 14:12:37.630389] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:33.228 [2024-09-30 14:12:37.630399] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:33.228 [2024-09-30 14:12:37.630415] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:13:33.228 [2024-09-30 14:12:37.630443] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:33.228 pt1 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.228 "name": "raid_bdev1", 00:13:33.228 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:33.228 "strip_size_kb": 64, 00:13:33.228 "state": "configuring", 00:13:33.228 "raid_level": "raid5f", 00:13:33.228 "superblock": true, 00:13:33.228 "num_base_bdevs": 3, 00:13:33.228 "num_base_bdevs_discovered": 1, 00:13:33.228 "num_base_bdevs_operational": 2, 00:13:33.228 "base_bdevs_list": [ 00:13:33.228 { 00:13:33.228 "name": null, 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.228 "is_configured": false, 00:13:33.228 "data_offset": 2048, 00:13:33.228 "data_size": 63488 00:13:33.228 }, 00:13:33.228 { 00:13:33.228 "name": "pt2", 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:33.228 "is_configured": true, 00:13:33.228 "data_offset": 2048, 00:13:33.228 "data_size": 63488 00:13:33.228 }, 00:13:33.228 { 00:13:33.228 "name": null, 00:13:33.228 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:33.228 "is_configured": false, 00:13:33.228 "data_offset": 2048, 00:13:33.228 "data_size": 63488 00:13:33.228 } 00:13:33.228 ] 00:13:33.228 }' 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.228 14:12:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:33.488 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.488 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.488 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:33.488 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.748 [2024-09-30 14:12:38.151161] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:33.748 [2024-09-30 14:12:38.151281] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:33.748 [2024-09-30 14:12:38.151330] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:13:33.748 [2024-09-30 14:12:38.151360] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:33.748 [2024-09-30 14:12:38.151718] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:33.748 [2024-09-30 14:12:38.151775] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:33.748 [2024-09-30 14:12:38.151862] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:33.748 [2024-09-30 14:12:38.151909] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:33.748 [2024-09-30 14:12:38.152002] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:13:33.748 [2024-09-30 14:12:38.152039] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:33.748 [2024-09-30 14:12:38.152279] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:33.748 [2024-09-30 14:12:38.152751] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:13:33.748 [2024-09-30 14:12:38.152800] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:13:33.748 [2024-09-30 14:12:38.152969] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:33.748 pt3 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.748 "name": "raid_bdev1", 00:13:33.748 "uuid": "95b1da33-3157-48b7-88ae-d6faf9e4d59a", 00:13:33.748 "strip_size_kb": 64, 00:13:33.748 "state": "online", 00:13:33.748 "raid_level": "raid5f", 00:13:33.748 "superblock": true, 00:13:33.748 "num_base_bdevs": 3, 00:13:33.748 "num_base_bdevs_discovered": 2, 00:13:33.748 "num_base_bdevs_operational": 2, 00:13:33.748 "base_bdevs_list": [ 00:13:33.748 { 00:13:33.748 "name": null, 00:13:33.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.748 "is_configured": false, 00:13:33.748 "data_offset": 2048, 00:13:33.748 "data_size": 63488 00:13:33.748 }, 00:13:33.748 { 00:13:33.748 "name": "pt2", 00:13:33.748 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:33.748 "is_configured": true, 00:13:33.748 "data_offset": 2048, 00:13:33.748 "data_size": 63488 00:13:33.748 }, 00:13:33.748 { 00:13:33.748 "name": "pt3", 00:13:33.748 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:33.748 "is_configured": true, 00:13:33.748 "data_offset": 2048, 00:13:33.748 "data_size": 63488 00:13:33.748 } 00:13:33.748 ] 00:13:33.748 }' 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.748 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.008 [2024-09-30 14:12:38.638557] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 95b1da33-3157-48b7-88ae-d6faf9e4d59a '!=' 95b1da33-3157-48b7-88ae-d6faf9e4d59a ']' 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 92677 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 92677 ']' 00:13:34.008 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 92677 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 92677 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:34.269 killing process with pid 92677 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 92677' 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 92677 00:13:34.269 [2024-09-30 14:12:38.694073] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:34.269 [2024-09-30 14:12:38.694156] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:34.269 [2024-09-30 14:12:38.694202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:34.269 [2024-09-30 14:12:38.694213] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:13:34.269 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 92677 00:13:34.269 [2024-09-30 14:12:38.727334] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:34.537 ************************************ 00:13:34.537 END TEST raid5f_superblock_test 00:13:34.537 ************************************ 00:13:34.537 14:12:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:34.537 00:13:34.537 real 0m6.524s 00:13:34.537 user 0m10.854s 00:13:34.537 sys 0m1.454s 00:13:34.537 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:34.537 14:12:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.537 14:12:39 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:34.537 14:12:39 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:13:34.537 14:12:39 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:34.537 14:12:39 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:34.538 14:12:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:34.538 ************************************ 00:13:34.538 START TEST raid5f_rebuild_test 00:13:34.538 ************************************ 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 false false true 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=93110 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 93110 00:13:34.538 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 93110 ']' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:34.538 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.538 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:34.538 Zero copy mechanism will not be used. 00:13:34.538 [2024-09-30 14:12:39.160861] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:13:34.538 [2024-09-30 14:12:39.160973] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid93110 ] 00:13:34.798 [2024-09-30 14:12:39.291971] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:13:34.798 [2024-09-30 14:12:39.319091] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:34.798 [2024-09-30 14:12:39.363332] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:34.798 [2024-09-30 14:12:39.405037] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:34.798 [2024-09-30 14:12:39.405080] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.367 14:12:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.367 BaseBdev1_malloc 00:13:35.367 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.367 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:35.367 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.368 [2024-09-30 14:12:40.010646] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:35.368 [2024-09-30 14:12:40.010801] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:35.368 [2024-09-30 14:12:40.010838] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:35.368 [2024-09-30 14:12:40.010861] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:35.368 [2024-09-30 14:12:40.012998] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:35.368 [2024-09-30 14:12:40.013039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:35.368 BaseBdev1 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.368 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 BaseBdev2_malloc 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 [2024-09-30 14:12:40.055189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:35.628 [2024-09-30 14:12:40.055299] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:35.628 [2024-09-30 14:12:40.055341] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:35.628 [2024-09-30 14:12:40.055367] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:35.628 [2024-09-30 14:12:40.060286] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:35.628 [2024-09-30 14:12:40.060362] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:35.628 BaseBdev2 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 BaseBdev3_malloc 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 [2024-09-30 14:12:40.086650] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:35.628 [2024-09-30 14:12:40.086703] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:35.628 [2024-09-30 14:12:40.086721] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:35.628 [2024-09-30 14:12:40.086731] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:35.628 [2024-09-30 14:12:40.088680] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:35.628 [2024-09-30 14:12:40.088718] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:35.628 BaseBdev3 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 spare_malloc 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.628 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.628 spare_delay 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.629 [2024-09-30 14:12:40.127251] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:35.629 [2024-09-30 14:12:40.127301] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:35.629 [2024-09-30 14:12:40.127319] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:35.629 [2024-09-30 14:12:40.127329] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:35.629 [2024-09-30 14:12:40.129304] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:35.629 [2024-09-30 14:12:40.129341] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:35.629 spare 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.629 [2024-09-30 14:12:40.139326] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:35.629 [2024-09-30 14:12:40.141089] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:35.629 [2024-09-30 14:12:40.141143] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:35.629 [2024-09-30 14:12:40.141212] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:35.629 [2024-09-30 14:12:40.141220] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:35.629 [2024-09-30 14:12:40.141449] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:13:35.629 [2024-09-30 14:12:40.141840] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:35.629 [2024-09-30 14:12:40.141853] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:35.629 [2024-09-30 14:12:40.141963] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.629 "name": "raid_bdev1", 00:13:35.629 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:35.629 "strip_size_kb": 64, 00:13:35.629 "state": "online", 00:13:35.629 "raid_level": "raid5f", 00:13:35.629 "superblock": false, 00:13:35.629 "num_base_bdevs": 3, 00:13:35.629 "num_base_bdevs_discovered": 3, 00:13:35.629 "num_base_bdevs_operational": 3, 00:13:35.629 "base_bdevs_list": [ 00:13:35.629 { 00:13:35.629 "name": "BaseBdev1", 00:13:35.629 "uuid": "1e2ff362-ecfd-58e3-ba21-34df60a83eba", 00:13:35.629 "is_configured": true, 00:13:35.629 "data_offset": 0, 00:13:35.629 "data_size": 65536 00:13:35.629 }, 00:13:35.629 { 00:13:35.629 "name": "BaseBdev2", 00:13:35.629 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:35.629 "is_configured": true, 00:13:35.629 "data_offset": 0, 00:13:35.629 "data_size": 65536 00:13:35.629 }, 00:13:35.629 { 00:13:35.629 "name": "BaseBdev3", 00:13:35.629 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:35.629 "is_configured": true, 00:13:35.629 "data_offset": 0, 00:13:35.629 "data_size": 65536 00:13:35.629 } 00:13:35.629 ] 00:13:35.629 }' 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.629 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.198 [2024-09-30 14:12:40.558835] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:36.198 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:36.199 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:36.199 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:36.199 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.199 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:36.199 [2024-09-30 14:12:40.822328] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:36.199 /dev/nbd0 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:36.459 1+0 records in 00:13:36.459 1+0 records out 00:13:36.459 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000580692 s, 7.1 MB/s 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:36.459 14:12:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:36.718 512+0 records in 00:13:36.718 512+0 records out 00:13:36.718 67108864 bytes (67 MB, 64 MiB) copied, 0.322354 s, 208 MB/s 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:36.718 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:36.978 [2024-09-30 14:12:41.423501] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.978 [2024-09-30 14:12:41.441144] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.978 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.979 "name": "raid_bdev1", 00:13:36.979 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:36.979 "strip_size_kb": 64, 00:13:36.979 "state": "online", 00:13:36.979 "raid_level": "raid5f", 00:13:36.979 "superblock": false, 00:13:36.979 "num_base_bdevs": 3, 00:13:36.979 "num_base_bdevs_discovered": 2, 00:13:36.979 "num_base_bdevs_operational": 2, 00:13:36.979 "base_bdevs_list": [ 00:13:36.979 { 00:13:36.979 "name": null, 00:13:36.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.979 "is_configured": false, 00:13:36.979 "data_offset": 0, 00:13:36.979 "data_size": 65536 00:13:36.979 }, 00:13:36.979 { 00:13:36.979 "name": "BaseBdev2", 00:13:36.979 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:36.979 "is_configured": true, 00:13:36.979 "data_offset": 0, 00:13:36.979 "data_size": 65536 00:13:36.979 }, 00:13:36.979 { 00:13:36.979 "name": "BaseBdev3", 00:13:36.979 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:36.979 "is_configured": true, 00:13:36.979 "data_offset": 0, 00:13:36.979 "data_size": 65536 00:13:36.979 } 00:13:36.979 ] 00:13:36.979 }' 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.979 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.238 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:37.238 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.239 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.239 [2024-09-30 14:12:41.884472] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:37.239 [2024-09-30 14:12:41.888025] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:13:37.239 [2024-09-30 14:12:41.890079] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:37.239 14:12:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.239 14:12:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:38.619 "name": "raid_bdev1", 00:13:38.619 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:38.619 "strip_size_kb": 64, 00:13:38.619 "state": "online", 00:13:38.619 "raid_level": "raid5f", 00:13:38.619 "superblock": false, 00:13:38.619 "num_base_bdevs": 3, 00:13:38.619 "num_base_bdevs_discovered": 3, 00:13:38.619 "num_base_bdevs_operational": 3, 00:13:38.619 "process": { 00:13:38.619 "type": "rebuild", 00:13:38.619 "target": "spare", 00:13:38.619 "progress": { 00:13:38.619 "blocks": 20480, 00:13:38.619 "percent": 15 00:13:38.619 } 00:13:38.619 }, 00:13:38.619 "base_bdevs_list": [ 00:13:38.619 { 00:13:38.619 "name": "spare", 00:13:38.619 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:38.619 "is_configured": true, 00:13:38.619 "data_offset": 0, 00:13:38.619 "data_size": 65536 00:13:38.619 }, 00:13:38.619 { 00:13:38.619 "name": "BaseBdev2", 00:13:38.619 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:38.619 "is_configured": true, 00:13:38.619 "data_offset": 0, 00:13:38.619 "data_size": 65536 00:13:38.619 }, 00:13:38.619 { 00:13:38.619 "name": "BaseBdev3", 00:13:38.619 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:38.619 "is_configured": true, 00:13:38.619 "data_offset": 0, 00:13:38.619 "data_size": 65536 00:13:38.619 } 00:13:38.619 ] 00:13:38.619 }' 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:38.619 14:12:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.619 [2024-09-30 14:12:43.032611] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:38.619 [2024-09-30 14:12:43.096542] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:38.619 [2024-09-30 14:12:43.096632] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:38.619 [2024-09-30 14:12:43.096650] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:38.619 [2024-09-30 14:12:43.096661] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.619 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.619 "name": "raid_bdev1", 00:13:38.619 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:38.619 "strip_size_kb": 64, 00:13:38.619 "state": "online", 00:13:38.619 "raid_level": "raid5f", 00:13:38.619 "superblock": false, 00:13:38.619 "num_base_bdevs": 3, 00:13:38.619 "num_base_bdevs_discovered": 2, 00:13:38.619 "num_base_bdevs_operational": 2, 00:13:38.619 "base_bdevs_list": [ 00:13:38.619 { 00:13:38.619 "name": null, 00:13:38.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.619 "is_configured": false, 00:13:38.620 "data_offset": 0, 00:13:38.620 "data_size": 65536 00:13:38.620 }, 00:13:38.620 { 00:13:38.620 "name": "BaseBdev2", 00:13:38.620 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:38.620 "is_configured": true, 00:13:38.620 "data_offset": 0, 00:13:38.620 "data_size": 65536 00:13:38.620 }, 00:13:38.620 { 00:13:38.620 "name": "BaseBdev3", 00:13:38.620 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:38.620 "is_configured": true, 00:13:38.620 "data_offset": 0, 00:13:38.620 "data_size": 65536 00:13:38.620 } 00:13:38.620 ] 00:13:38.620 }' 00:13:38.620 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.620 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:39.190 "name": "raid_bdev1", 00:13:39.190 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:39.190 "strip_size_kb": 64, 00:13:39.190 "state": "online", 00:13:39.190 "raid_level": "raid5f", 00:13:39.190 "superblock": false, 00:13:39.190 "num_base_bdevs": 3, 00:13:39.190 "num_base_bdevs_discovered": 2, 00:13:39.190 "num_base_bdevs_operational": 2, 00:13:39.190 "base_bdevs_list": [ 00:13:39.190 { 00:13:39.190 "name": null, 00:13:39.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.190 "is_configured": false, 00:13:39.190 "data_offset": 0, 00:13:39.190 "data_size": 65536 00:13:39.190 }, 00:13:39.190 { 00:13:39.190 "name": "BaseBdev2", 00:13:39.190 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:39.190 "is_configured": true, 00:13:39.190 "data_offset": 0, 00:13:39.190 "data_size": 65536 00:13:39.190 }, 00:13:39.190 { 00:13:39.190 "name": "BaseBdev3", 00:13:39.190 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:39.190 "is_configured": true, 00:13:39.190 "data_offset": 0, 00:13:39.190 "data_size": 65536 00:13:39.190 } 00:13:39.190 ] 00:13:39.190 }' 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.190 [2024-09-30 14:12:43.712636] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:39.190 [2024-09-30 14:12:43.715677] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b820 00:13:39.190 [2024-09-30 14:12:43.717725] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.190 14:12:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:40.129 "name": "raid_bdev1", 00:13:40.129 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:40.129 "strip_size_kb": 64, 00:13:40.129 "state": "online", 00:13:40.129 "raid_level": "raid5f", 00:13:40.129 "superblock": false, 00:13:40.129 "num_base_bdevs": 3, 00:13:40.129 "num_base_bdevs_discovered": 3, 00:13:40.129 "num_base_bdevs_operational": 3, 00:13:40.129 "process": { 00:13:40.129 "type": "rebuild", 00:13:40.129 "target": "spare", 00:13:40.129 "progress": { 00:13:40.129 "blocks": 20480, 00:13:40.129 "percent": 15 00:13:40.129 } 00:13:40.129 }, 00:13:40.129 "base_bdevs_list": [ 00:13:40.129 { 00:13:40.129 "name": "spare", 00:13:40.129 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:40.129 "is_configured": true, 00:13:40.129 "data_offset": 0, 00:13:40.129 "data_size": 65536 00:13:40.129 }, 00:13:40.129 { 00:13:40.129 "name": "BaseBdev2", 00:13:40.129 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:40.129 "is_configured": true, 00:13:40.129 "data_offset": 0, 00:13:40.129 "data_size": 65536 00:13:40.129 }, 00:13:40.129 { 00:13:40.129 "name": "BaseBdev3", 00:13:40.129 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:40.129 "is_configured": true, 00:13:40.129 "data_offset": 0, 00:13:40.129 "data_size": 65536 00:13:40.129 } 00:13:40.129 ] 00:13:40.129 }' 00:13:40.129 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=445 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.389 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:40.389 "name": "raid_bdev1", 00:13:40.389 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:40.389 "strip_size_kb": 64, 00:13:40.389 "state": "online", 00:13:40.389 "raid_level": "raid5f", 00:13:40.389 "superblock": false, 00:13:40.389 "num_base_bdevs": 3, 00:13:40.389 "num_base_bdevs_discovered": 3, 00:13:40.389 "num_base_bdevs_operational": 3, 00:13:40.389 "process": { 00:13:40.389 "type": "rebuild", 00:13:40.389 "target": "spare", 00:13:40.389 "progress": { 00:13:40.389 "blocks": 22528, 00:13:40.389 "percent": 17 00:13:40.389 } 00:13:40.390 }, 00:13:40.390 "base_bdevs_list": [ 00:13:40.390 { 00:13:40.390 "name": "spare", 00:13:40.390 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev2", 00:13:40.390 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev3", 00:13:40.390 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 } 00:13:40.390 ] 00:13:40.390 }' 00:13:40.390 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.390 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:40.390 14:12:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.390 14:12:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:40.390 14:12:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:41.771 "name": "raid_bdev1", 00:13:41.771 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:41.771 "strip_size_kb": 64, 00:13:41.771 "state": "online", 00:13:41.771 "raid_level": "raid5f", 00:13:41.771 "superblock": false, 00:13:41.771 "num_base_bdevs": 3, 00:13:41.771 "num_base_bdevs_discovered": 3, 00:13:41.771 "num_base_bdevs_operational": 3, 00:13:41.771 "process": { 00:13:41.771 "type": "rebuild", 00:13:41.771 "target": "spare", 00:13:41.771 "progress": { 00:13:41.771 "blocks": 47104, 00:13:41.771 "percent": 35 00:13:41.771 } 00:13:41.771 }, 00:13:41.771 "base_bdevs_list": [ 00:13:41.771 { 00:13:41.771 "name": "spare", 00:13:41.771 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:41.771 "is_configured": true, 00:13:41.771 "data_offset": 0, 00:13:41.771 "data_size": 65536 00:13:41.771 }, 00:13:41.771 { 00:13:41.771 "name": "BaseBdev2", 00:13:41.771 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:41.771 "is_configured": true, 00:13:41.771 "data_offset": 0, 00:13:41.771 "data_size": 65536 00:13:41.771 }, 00:13:41.771 { 00:13:41.771 "name": "BaseBdev3", 00:13:41.771 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:41.771 "is_configured": true, 00:13:41.771 "data_offset": 0, 00:13:41.771 "data_size": 65536 00:13:41.771 } 00:13:41.771 ] 00:13:41.771 }' 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:41.771 14:12:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:42.710 "name": "raid_bdev1", 00:13:42.710 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:42.710 "strip_size_kb": 64, 00:13:42.710 "state": "online", 00:13:42.710 "raid_level": "raid5f", 00:13:42.710 "superblock": false, 00:13:42.710 "num_base_bdevs": 3, 00:13:42.710 "num_base_bdevs_discovered": 3, 00:13:42.710 "num_base_bdevs_operational": 3, 00:13:42.710 "process": { 00:13:42.710 "type": "rebuild", 00:13:42.710 "target": "spare", 00:13:42.710 "progress": { 00:13:42.710 "blocks": 69632, 00:13:42.710 "percent": 53 00:13:42.710 } 00:13:42.710 }, 00:13:42.710 "base_bdevs_list": [ 00:13:42.710 { 00:13:42.710 "name": "spare", 00:13:42.710 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:42.710 "is_configured": true, 00:13:42.710 "data_offset": 0, 00:13:42.710 "data_size": 65536 00:13:42.710 }, 00:13:42.710 { 00:13:42.710 "name": "BaseBdev2", 00:13:42.710 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:42.710 "is_configured": true, 00:13:42.710 "data_offset": 0, 00:13:42.710 "data_size": 65536 00:13:42.710 }, 00:13:42.710 { 00:13:42.710 "name": "BaseBdev3", 00:13:42.710 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:42.710 "is_configured": true, 00:13:42.710 "data_offset": 0, 00:13:42.710 "data_size": 65536 00:13:42.710 } 00:13:42.710 ] 00:13:42.710 }' 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:42.710 14:12:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.091 "name": "raid_bdev1", 00:13:44.091 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:44.091 "strip_size_kb": 64, 00:13:44.091 "state": "online", 00:13:44.091 "raid_level": "raid5f", 00:13:44.091 "superblock": false, 00:13:44.091 "num_base_bdevs": 3, 00:13:44.091 "num_base_bdevs_discovered": 3, 00:13:44.091 "num_base_bdevs_operational": 3, 00:13:44.091 "process": { 00:13:44.091 "type": "rebuild", 00:13:44.091 "target": "spare", 00:13:44.091 "progress": { 00:13:44.091 "blocks": 92160, 00:13:44.091 "percent": 70 00:13:44.091 } 00:13:44.091 }, 00:13:44.091 "base_bdevs_list": [ 00:13:44.091 { 00:13:44.091 "name": "spare", 00:13:44.091 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:44.091 "is_configured": true, 00:13:44.091 "data_offset": 0, 00:13:44.091 "data_size": 65536 00:13:44.091 }, 00:13:44.091 { 00:13:44.091 "name": "BaseBdev2", 00:13:44.091 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:44.091 "is_configured": true, 00:13:44.091 "data_offset": 0, 00:13:44.091 "data_size": 65536 00:13:44.091 }, 00:13:44.091 { 00:13:44.091 "name": "BaseBdev3", 00:13:44.091 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:44.091 "is_configured": true, 00:13:44.091 "data_offset": 0, 00:13:44.091 "data_size": 65536 00:13:44.091 } 00:13:44.091 ] 00:13:44.091 }' 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:44.091 14:12:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:45.030 "name": "raid_bdev1", 00:13:45.030 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:45.030 "strip_size_kb": 64, 00:13:45.030 "state": "online", 00:13:45.030 "raid_level": "raid5f", 00:13:45.030 "superblock": false, 00:13:45.030 "num_base_bdevs": 3, 00:13:45.030 "num_base_bdevs_discovered": 3, 00:13:45.030 "num_base_bdevs_operational": 3, 00:13:45.030 "process": { 00:13:45.030 "type": "rebuild", 00:13:45.030 "target": "spare", 00:13:45.030 "progress": { 00:13:45.030 "blocks": 116736, 00:13:45.030 "percent": 89 00:13:45.030 } 00:13:45.030 }, 00:13:45.030 "base_bdevs_list": [ 00:13:45.030 { 00:13:45.030 "name": "spare", 00:13:45.030 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:45.030 "is_configured": true, 00:13:45.030 "data_offset": 0, 00:13:45.030 "data_size": 65536 00:13:45.030 }, 00:13:45.030 { 00:13:45.030 "name": "BaseBdev2", 00:13:45.030 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:45.030 "is_configured": true, 00:13:45.030 "data_offset": 0, 00:13:45.030 "data_size": 65536 00:13:45.030 }, 00:13:45.030 { 00:13:45.030 "name": "BaseBdev3", 00:13:45.030 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:45.030 "is_configured": true, 00:13:45.030 "data_offset": 0, 00:13:45.030 "data_size": 65536 00:13:45.030 } 00:13:45.030 ] 00:13:45.030 }' 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:45.030 14:12:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:45.600 [2024-09-30 14:12:50.149910] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:45.600 [2024-09-30 14:12:50.150032] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:45.600 [2024-09-30 14:12:50.150073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:46.170 "name": "raid_bdev1", 00:13:46.170 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:46.170 "strip_size_kb": 64, 00:13:46.170 "state": "online", 00:13:46.170 "raid_level": "raid5f", 00:13:46.170 "superblock": false, 00:13:46.170 "num_base_bdevs": 3, 00:13:46.170 "num_base_bdevs_discovered": 3, 00:13:46.170 "num_base_bdevs_operational": 3, 00:13:46.170 "base_bdevs_list": [ 00:13:46.170 { 00:13:46.170 "name": "spare", 00:13:46.170 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:46.170 "is_configured": true, 00:13:46.170 "data_offset": 0, 00:13:46.170 "data_size": 65536 00:13:46.170 }, 00:13:46.170 { 00:13:46.170 "name": "BaseBdev2", 00:13:46.170 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:46.170 "is_configured": true, 00:13:46.170 "data_offset": 0, 00:13:46.170 "data_size": 65536 00:13:46.170 }, 00:13:46.170 { 00:13:46.170 "name": "BaseBdev3", 00:13:46.170 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:46.170 "is_configured": true, 00:13:46.170 "data_offset": 0, 00:13:46.170 "data_size": 65536 00:13:46.170 } 00:13:46.170 ] 00:13:46.170 }' 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.170 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:46.430 "name": "raid_bdev1", 00:13:46.430 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:46.430 "strip_size_kb": 64, 00:13:46.430 "state": "online", 00:13:46.430 "raid_level": "raid5f", 00:13:46.430 "superblock": false, 00:13:46.430 "num_base_bdevs": 3, 00:13:46.430 "num_base_bdevs_discovered": 3, 00:13:46.430 "num_base_bdevs_operational": 3, 00:13:46.430 "base_bdevs_list": [ 00:13:46.430 { 00:13:46.430 "name": "spare", 00:13:46.430 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 }, 00:13:46.430 { 00:13:46.430 "name": "BaseBdev2", 00:13:46.430 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 }, 00:13:46.430 { 00:13:46.430 "name": "BaseBdev3", 00:13:46.430 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 } 00:13:46.430 ] 00:13:46.430 }' 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.430 "name": "raid_bdev1", 00:13:46.430 "uuid": "d031a6b4-0b8c-42d5-b72e-ea1618762775", 00:13:46.430 "strip_size_kb": 64, 00:13:46.430 "state": "online", 00:13:46.430 "raid_level": "raid5f", 00:13:46.430 "superblock": false, 00:13:46.430 "num_base_bdevs": 3, 00:13:46.430 "num_base_bdevs_discovered": 3, 00:13:46.430 "num_base_bdevs_operational": 3, 00:13:46.430 "base_bdevs_list": [ 00:13:46.430 { 00:13:46.430 "name": "spare", 00:13:46.430 "uuid": "b9b0f504-f83f-5050-a3b6-03e2c081e73f", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 }, 00:13:46.430 { 00:13:46.430 "name": "BaseBdev2", 00:13:46.430 "uuid": "d3507d52-34f9-5e0c-aab3-810b430aba2c", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 }, 00:13:46.430 { 00:13:46.430 "name": "BaseBdev3", 00:13:46.430 "uuid": "69623746-9144-5a47-86c0-764b2fa13eb4", 00:13:46.430 "is_configured": true, 00:13:46.430 "data_offset": 0, 00:13:46.430 "data_size": 65536 00:13:46.430 } 00:13:46.430 ] 00:13:46.430 }' 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.430 14:12:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.999 [2024-09-30 14:12:51.408722] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:46.999 [2024-09-30 14:12:51.408799] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:46.999 [2024-09-30 14:12:51.408872] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:46.999 [2024-09-30 14:12:51.408938] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:46.999 [2024-09-30 14:12:51.408949] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:46.999 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:47.259 /dev/nbd0 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:47.259 1+0 records in 00:13:47.259 1+0 records out 00:13:47.259 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000387833 s, 10.6 MB/s 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:47.259 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:47.519 /dev/nbd1 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:47.519 1+0 records in 00:13:47.519 1+0 records out 00:13:47.519 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000342243 s, 12.0 MB/s 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:47.519 14:12:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:47.519 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:47.778 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 93110 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 93110 ']' 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 93110 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 93110 00:13:48.038 killing process with pid 93110 00:13:48.038 Received shutdown signal, test time was about 60.000000 seconds 00:13:48.038 00:13:48.038 Latency(us) 00:13:48.038 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:48.038 =================================================================================================================== 00:13:48.038 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 93110' 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 93110 00:13:48.038 [2024-09-30 14:12:52.545535] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:48.038 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 93110 00:13:48.038 [2024-09-30 14:12:52.586249] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:48.298 00:13:48.298 real 0m13.759s 00:13:48.298 user 0m17.205s 00:13:48.298 sys 0m2.088s 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.298 ************************************ 00:13:48.298 END TEST raid5f_rebuild_test 00:13:48.298 ************************************ 00:13:48.298 14:12:52 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:13:48.298 14:12:52 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:48.298 14:12:52 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:48.298 14:12:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:48.298 ************************************ 00:13:48.298 START TEST raid5f_rebuild_test_sb 00:13:48.298 ************************************ 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 true false true 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=93533 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 93533 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 93533 ']' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:48.298 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:48.298 14:12:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.557 [2024-09-30 14:12:53.004362] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:13:48.557 [2024-09-30 14:12:53.004533] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid93533 ] 00:13:48.557 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:48.557 Zero copy mechanism will not be used. 00:13:48.557 [2024-09-30 14:12:53.142634] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:13:48.557 [2024-09-30 14:12:53.170851] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:48.817 [2024-09-30 14:12:53.219746] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:48.817 [2024-09-30 14:12:53.262534] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:48.817 [2024-09-30 14:12:53.262576] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 BaseBdev1_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 [2024-09-30 14:12:53.836240] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:49.387 [2024-09-30 14:12:53.836306] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:49.387 [2024-09-30 14:12:53.836329] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:49.387 [2024-09-30 14:12:53.836346] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:49.387 [2024-09-30 14:12:53.838321] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:49.387 [2024-09-30 14:12:53.838359] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:49.387 BaseBdev1 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 BaseBdev2_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 [2024-09-30 14:12:53.878100] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:49.387 [2024-09-30 14:12:53.878186] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:49.387 [2024-09-30 14:12:53.878217] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:49.387 [2024-09-30 14:12:53.878237] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:49.387 [2024-09-30 14:12:53.881978] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:49.387 [2024-09-30 14:12:53.882037] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:49.387 BaseBdev2 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 BaseBdev3_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 [2024-09-30 14:12:53.906867] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:49.387 [2024-09-30 14:12:53.906914] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:49.387 [2024-09-30 14:12:53.906933] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:49.387 [2024-09-30 14:12:53.906943] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:49.387 [2024-09-30 14:12:53.908922] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:49.387 [2024-09-30 14:12:53.908960] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:49.387 BaseBdev3 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 spare_malloc 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 spare_delay 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 [2024-09-30 14:12:53.947172] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:49.387 [2024-09-30 14:12:53.947227] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:49.387 [2024-09-30 14:12:53.947247] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:49.387 [2024-09-30 14:12:53.947257] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:49.387 [2024-09-30 14:12:53.949214] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:49.387 [2024-09-30 14:12:53.949252] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:49.387 spare 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.387 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.387 [2024-09-30 14:12:53.959241] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:49.387 [2024-09-30 14:12:53.960905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:49.387 [2024-09-30 14:12:53.960960] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:49.387 [2024-09-30 14:12:53.961096] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:49.387 [2024-09-30 14:12:53.961113] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:49.387 [2024-09-30 14:12:53.961318] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:13:49.387 [2024-09-30 14:12:53.961714] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:49.388 [2024-09-30 14:12:53.961734] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:49.388 [2024-09-30 14:12:53.961837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.388 14:12:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.388 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.388 "name": "raid_bdev1", 00:13:49.388 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:49.388 "strip_size_kb": 64, 00:13:49.388 "state": "online", 00:13:49.388 "raid_level": "raid5f", 00:13:49.388 "superblock": true, 00:13:49.388 "num_base_bdevs": 3, 00:13:49.388 "num_base_bdevs_discovered": 3, 00:13:49.388 "num_base_bdevs_operational": 3, 00:13:49.388 "base_bdevs_list": [ 00:13:49.388 { 00:13:49.388 "name": "BaseBdev1", 00:13:49.388 "uuid": "3e95d9c6-a20d-5f99-beda-ff8bdbe41ccb", 00:13:49.388 "is_configured": true, 00:13:49.388 "data_offset": 2048, 00:13:49.388 "data_size": 63488 00:13:49.388 }, 00:13:49.388 { 00:13:49.388 "name": "BaseBdev2", 00:13:49.388 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:49.388 "is_configured": true, 00:13:49.388 "data_offset": 2048, 00:13:49.388 "data_size": 63488 00:13:49.388 }, 00:13:49.388 { 00:13:49.388 "name": "BaseBdev3", 00:13:49.388 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:49.388 "is_configured": true, 00:13:49.388 "data_offset": 2048, 00:13:49.388 "data_size": 63488 00:13:49.388 } 00:13:49.388 ] 00:13:49.388 }' 00:13:49.388 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.388 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:50.006 [2024-09-30 14:12:54.366702] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:50.006 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:50.287 [2024-09-30 14:12:54.646101] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:50.287 /dev/nbd0 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:50.287 1+0 records in 00:13:50.287 1+0 records out 00:13:50.287 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000203465 s, 20.1 MB/s 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:50.287 14:12:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:13:50.547 496+0 records in 00:13:50.547 496+0 records out 00:13:50.547 65011712 bytes (65 MB, 62 MiB) copied, 0.295722 s, 220 MB/s 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:50.547 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:50.547 [2024-09-30 14:12:55.190961] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.807 [2024-09-30 14:12:55.234936] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.807 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.807 "name": "raid_bdev1", 00:13:50.807 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:50.807 "strip_size_kb": 64, 00:13:50.807 "state": "online", 00:13:50.807 "raid_level": "raid5f", 00:13:50.807 "superblock": true, 00:13:50.807 "num_base_bdevs": 3, 00:13:50.807 "num_base_bdevs_discovered": 2, 00:13:50.807 "num_base_bdevs_operational": 2, 00:13:50.807 "base_bdevs_list": [ 00:13:50.807 { 00:13:50.807 "name": null, 00:13:50.807 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.807 "is_configured": false, 00:13:50.807 "data_offset": 0, 00:13:50.807 "data_size": 63488 00:13:50.807 }, 00:13:50.807 { 00:13:50.807 "name": "BaseBdev2", 00:13:50.807 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:50.807 "is_configured": true, 00:13:50.807 "data_offset": 2048, 00:13:50.807 "data_size": 63488 00:13:50.808 }, 00:13:50.808 { 00:13:50.808 "name": "BaseBdev3", 00:13:50.808 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:50.808 "is_configured": true, 00:13:50.808 "data_offset": 2048, 00:13:50.808 "data_size": 63488 00:13:50.808 } 00:13:50.808 ] 00:13:50.808 }' 00:13:50.808 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.808 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.066 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:51.066 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.066 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.066 [2024-09-30 14:12:55.710157] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:51.066 [2024-09-30 14:12:55.713887] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000029050 00:13:51.066 [2024-09-30 14:12:55.716046] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:51.066 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.066 14:12:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:52.447 "name": "raid_bdev1", 00:13:52.447 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:52.447 "strip_size_kb": 64, 00:13:52.447 "state": "online", 00:13:52.447 "raid_level": "raid5f", 00:13:52.447 "superblock": true, 00:13:52.447 "num_base_bdevs": 3, 00:13:52.447 "num_base_bdevs_discovered": 3, 00:13:52.447 "num_base_bdevs_operational": 3, 00:13:52.447 "process": { 00:13:52.447 "type": "rebuild", 00:13:52.447 "target": "spare", 00:13:52.447 "progress": { 00:13:52.447 "blocks": 20480, 00:13:52.447 "percent": 16 00:13:52.447 } 00:13:52.447 }, 00:13:52.447 "base_bdevs_list": [ 00:13:52.447 { 00:13:52.447 "name": "spare", 00:13:52.447 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:52.447 "is_configured": true, 00:13:52.447 "data_offset": 2048, 00:13:52.447 "data_size": 63488 00:13:52.447 }, 00:13:52.447 { 00:13:52.447 "name": "BaseBdev2", 00:13:52.447 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:52.447 "is_configured": true, 00:13:52.447 "data_offset": 2048, 00:13:52.447 "data_size": 63488 00:13:52.447 }, 00:13:52.447 { 00:13:52.447 "name": "BaseBdev3", 00:13:52.447 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:52.447 "is_configured": true, 00:13:52.447 "data_offset": 2048, 00:13:52.447 "data_size": 63488 00:13:52.447 } 00:13:52.447 ] 00:13:52.447 }' 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.447 [2024-09-30 14:12:56.870894] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:52.447 [2024-09-30 14:12:56.922578] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:52.447 [2024-09-30 14:12:56.922635] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:52.447 [2024-09-30 14:12:56.922652] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:52.447 [2024-09-30 14:12:56.922659] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.447 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.447 "name": "raid_bdev1", 00:13:52.447 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:52.447 "strip_size_kb": 64, 00:13:52.447 "state": "online", 00:13:52.447 "raid_level": "raid5f", 00:13:52.447 "superblock": true, 00:13:52.447 "num_base_bdevs": 3, 00:13:52.447 "num_base_bdevs_discovered": 2, 00:13:52.447 "num_base_bdevs_operational": 2, 00:13:52.447 "base_bdevs_list": [ 00:13:52.447 { 00:13:52.448 "name": null, 00:13:52.448 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.448 "is_configured": false, 00:13:52.448 "data_offset": 0, 00:13:52.448 "data_size": 63488 00:13:52.448 }, 00:13:52.448 { 00:13:52.448 "name": "BaseBdev2", 00:13:52.448 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:52.448 "is_configured": true, 00:13:52.448 "data_offset": 2048, 00:13:52.448 "data_size": 63488 00:13:52.448 }, 00:13:52.448 { 00:13:52.448 "name": "BaseBdev3", 00:13:52.448 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:52.448 "is_configured": true, 00:13:52.448 "data_offset": 2048, 00:13:52.448 "data_size": 63488 00:13:52.448 } 00:13:52.448 ] 00:13:52.448 }' 00:13:52.448 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.448 14:12:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.016 "name": "raid_bdev1", 00:13:53.016 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:53.016 "strip_size_kb": 64, 00:13:53.016 "state": "online", 00:13:53.016 "raid_level": "raid5f", 00:13:53.016 "superblock": true, 00:13:53.016 "num_base_bdevs": 3, 00:13:53.016 "num_base_bdevs_discovered": 2, 00:13:53.016 "num_base_bdevs_operational": 2, 00:13:53.016 "base_bdevs_list": [ 00:13:53.016 { 00:13:53.016 "name": null, 00:13:53.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.016 "is_configured": false, 00:13:53.016 "data_offset": 0, 00:13:53.016 "data_size": 63488 00:13:53.016 }, 00:13:53.016 { 00:13:53.016 "name": "BaseBdev2", 00:13:53.016 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:53.016 "is_configured": true, 00:13:53.016 "data_offset": 2048, 00:13:53.016 "data_size": 63488 00:13:53.016 }, 00:13:53.016 { 00:13:53.016 "name": "BaseBdev3", 00:13:53.016 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:53.016 "is_configured": true, 00:13:53.016 "data_offset": 2048, 00:13:53.016 "data_size": 63488 00:13:53.016 } 00:13:53.016 ] 00:13:53.016 }' 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.016 [2024-09-30 14:12:57.539128] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:53.016 [2024-09-30 14:12:57.542616] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000029120 00:13:53.016 [2024-09-30 14:12:57.544676] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.016 14:12:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.953 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.954 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.954 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.954 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.954 "name": "raid_bdev1", 00:13:53.954 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:53.954 "strip_size_kb": 64, 00:13:53.954 "state": "online", 00:13:53.954 "raid_level": "raid5f", 00:13:53.954 "superblock": true, 00:13:53.954 "num_base_bdevs": 3, 00:13:53.954 "num_base_bdevs_discovered": 3, 00:13:53.954 "num_base_bdevs_operational": 3, 00:13:53.954 "process": { 00:13:53.954 "type": "rebuild", 00:13:53.954 "target": "spare", 00:13:53.954 "progress": { 00:13:53.954 "blocks": 20480, 00:13:53.954 "percent": 16 00:13:53.954 } 00:13:53.954 }, 00:13:53.954 "base_bdevs_list": [ 00:13:53.954 { 00:13:53.954 "name": "spare", 00:13:53.954 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:53.954 "is_configured": true, 00:13:53.954 "data_offset": 2048, 00:13:53.954 "data_size": 63488 00:13:53.954 }, 00:13:53.954 { 00:13:53.954 "name": "BaseBdev2", 00:13:53.954 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:53.954 "is_configured": true, 00:13:53.954 "data_offset": 2048, 00:13:53.954 "data_size": 63488 00:13:53.954 }, 00:13:53.954 { 00:13:53.954 "name": "BaseBdev3", 00:13:53.954 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:53.954 "is_configured": true, 00:13:53.954 "data_offset": 2048, 00:13:53.954 "data_size": 63488 00:13:53.954 } 00:13:53.954 ] 00:13:53.954 }' 00:13:53.954 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:54.213 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=459 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:54.213 "name": "raid_bdev1", 00:13:54.213 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:54.213 "strip_size_kb": 64, 00:13:54.213 "state": "online", 00:13:54.213 "raid_level": "raid5f", 00:13:54.213 "superblock": true, 00:13:54.213 "num_base_bdevs": 3, 00:13:54.213 "num_base_bdevs_discovered": 3, 00:13:54.213 "num_base_bdevs_operational": 3, 00:13:54.213 "process": { 00:13:54.213 "type": "rebuild", 00:13:54.213 "target": "spare", 00:13:54.213 "progress": { 00:13:54.213 "blocks": 22528, 00:13:54.213 "percent": 17 00:13:54.213 } 00:13:54.213 }, 00:13:54.213 "base_bdevs_list": [ 00:13:54.213 { 00:13:54.213 "name": "spare", 00:13:54.213 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:54.213 "is_configured": true, 00:13:54.213 "data_offset": 2048, 00:13:54.213 "data_size": 63488 00:13:54.213 }, 00:13:54.213 { 00:13:54.213 "name": "BaseBdev2", 00:13:54.213 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:54.213 "is_configured": true, 00:13:54.213 "data_offset": 2048, 00:13:54.213 "data_size": 63488 00:13:54.213 }, 00:13:54.213 { 00:13:54.213 "name": "BaseBdev3", 00:13:54.213 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:54.213 "is_configured": true, 00:13:54.213 "data_offset": 2048, 00:13:54.213 "data_size": 63488 00:13:54.213 } 00:13:54.213 ] 00:13:54.213 }' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:54.213 14:12:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:55.592 "name": "raid_bdev1", 00:13:55.592 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:55.592 "strip_size_kb": 64, 00:13:55.592 "state": "online", 00:13:55.592 "raid_level": "raid5f", 00:13:55.592 "superblock": true, 00:13:55.592 "num_base_bdevs": 3, 00:13:55.592 "num_base_bdevs_discovered": 3, 00:13:55.592 "num_base_bdevs_operational": 3, 00:13:55.592 "process": { 00:13:55.592 "type": "rebuild", 00:13:55.592 "target": "spare", 00:13:55.592 "progress": { 00:13:55.592 "blocks": 45056, 00:13:55.592 "percent": 35 00:13:55.592 } 00:13:55.592 }, 00:13:55.592 "base_bdevs_list": [ 00:13:55.592 { 00:13:55.592 "name": "spare", 00:13:55.592 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:55.592 "is_configured": true, 00:13:55.592 "data_offset": 2048, 00:13:55.592 "data_size": 63488 00:13:55.592 }, 00:13:55.592 { 00:13:55.592 "name": "BaseBdev2", 00:13:55.592 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:55.592 "is_configured": true, 00:13:55.592 "data_offset": 2048, 00:13:55.592 "data_size": 63488 00:13:55.592 }, 00:13:55.592 { 00:13:55.592 "name": "BaseBdev3", 00:13:55.592 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:55.592 "is_configured": true, 00:13:55.592 "data_offset": 2048, 00:13:55.592 "data_size": 63488 00:13:55.592 } 00:13:55.592 ] 00:13:55.592 }' 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:55.592 14:12:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.534 14:13:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:56.534 "name": "raid_bdev1", 00:13:56.534 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:56.534 "strip_size_kb": 64, 00:13:56.534 "state": "online", 00:13:56.534 "raid_level": "raid5f", 00:13:56.534 "superblock": true, 00:13:56.534 "num_base_bdevs": 3, 00:13:56.534 "num_base_bdevs_discovered": 3, 00:13:56.534 "num_base_bdevs_operational": 3, 00:13:56.534 "process": { 00:13:56.534 "type": "rebuild", 00:13:56.534 "target": "spare", 00:13:56.534 "progress": { 00:13:56.534 "blocks": 69632, 00:13:56.534 "percent": 54 00:13:56.534 } 00:13:56.534 }, 00:13:56.534 "base_bdevs_list": [ 00:13:56.534 { 00:13:56.534 "name": "spare", 00:13:56.534 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:56.534 "is_configured": true, 00:13:56.534 "data_offset": 2048, 00:13:56.534 "data_size": 63488 00:13:56.534 }, 00:13:56.534 { 00:13:56.534 "name": "BaseBdev2", 00:13:56.534 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:56.534 "is_configured": true, 00:13:56.534 "data_offset": 2048, 00:13:56.534 "data_size": 63488 00:13:56.534 }, 00:13:56.534 { 00:13:56.534 "name": "BaseBdev3", 00:13:56.534 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:56.534 "is_configured": true, 00:13:56.534 "data_offset": 2048, 00:13:56.534 "data_size": 63488 00:13:56.534 } 00:13:56.534 ] 00:13:56.534 }' 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:56.534 14:13:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.914 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:57.915 "name": "raid_bdev1", 00:13:57.915 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:57.915 "strip_size_kb": 64, 00:13:57.915 "state": "online", 00:13:57.915 "raid_level": "raid5f", 00:13:57.915 "superblock": true, 00:13:57.915 "num_base_bdevs": 3, 00:13:57.915 "num_base_bdevs_discovered": 3, 00:13:57.915 "num_base_bdevs_operational": 3, 00:13:57.915 "process": { 00:13:57.915 "type": "rebuild", 00:13:57.915 "target": "spare", 00:13:57.915 "progress": { 00:13:57.915 "blocks": 92160, 00:13:57.915 "percent": 72 00:13:57.915 } 00:13:57.915 }, 00:13:57.915 "base_bdevs_list": [ 00:13:57.915 { 00:13:57.915 "name": "spare", 00:13:57.915 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 }, 00:13:57.915 { 00:13:57.915 "name": "BaseBdev2", 00:13:57.915 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 }, 00:13:57.915 { 00:13:57.915 "name": "BaseBdev3", 00:13:57.915 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 } 00:13:57.915 ] 00:13:57.915 }' 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:57.915 14:13:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.852 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:58.852 "name": "raid_bdev1", 00:13:58.852 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:58.852 "strip_size_kb": 64, 00:13:58.852 "state": "online", 00:13:58.852 "raid_level": "raid5f", 00:13:58.852 "superblock": true, 00:13:58.852 "num_base_bdevs": 3, 00:13:58.852 "num_base_bdevs_discovered": 3, 00:13:58.852 "num_base_bdevs_operational": 3, 00:13:58.852 "process": { 00:13:58.852 "type": "rebuild", 00:13:58.852 "target": "spare", 00:13:58.852 "progress": { 00:13:58.852 "blocks": 116736, 00:13:58.852 "percent": 91 00:13:58.852 } 00:13:58.852 }, 00:13:58.852 "base_bdevs_list": [ 00:13:58.852 { 00:13:58.853 "name": "spare", 00:13:58.853 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:58.853 "is_configured": true, 00:13:58.853 "data_offset": 2048, 00:13:58.853 "data_size": 63488 00:13:58.853 }, 00:13:58.853 { 00:13:58.853 "name": "BaseBdev2", 00:13:58.853 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:58.853 "is_configured": true, 00:13:58.853 "data_offset": 2048, 00:13:58.853 "data_size": 63488 00:13:58.853 }, 00:13:58.853 { 00:13:58.853 "name": "BaseBdev3", 00:13:58.853 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:58.853 "is_configured": true, 00:13:58.853 "data_offset": 2048, 00:13:58.853 "data_size": 63488 00:13:58.853 } 00:13:58.853 ] 00:13:58.853 }' 00:13:58.853 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:58.853 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:58.853 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:58.853 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:58.853 14:13:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:59.421 [2024-09-30 14:13:03.775789] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:59.421 [2024-09-30 14:13:03.775894] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:59.421 [2024-09-30 14:13:03.776025] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:59.990 "name": "raid_bdev1", 00:13:59.990 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:13:59.990 "strip_size_kb": 64, 00:13:59.990 "state": "online", 00:13:59.990 "raid_level": "raid5f", 00:13:59.990 "superblock": true, 00:13:59.990 "num_base_bdevs": 3, 00:13:59.990 "num_base_bdevs_discovered": 3, 00:13:59.990 "num_base_bdevs_operational": 3, 00:13:59.990 "base_bdevs_list": [ 00:13:59.990 { 00:13:59.990 "name": "spare", 00:13:59.990 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:13:59.990 "is_configured": true, 00:13:59.990 "data_offset": 2048, 00:13:59.990 "data_size": 63488 00:13:59.990 }, 00:13:59.990 { 00:13:59.990 "name": "BaseBdev2", 00:13:59.990 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:13:59.990 "is_configured": true, 00:13:59.990 "data_offset": 2048, 00:13:59.990 "data_size": 63488 00:13:59.990 }, 00:13:59.990 { 00:13:59.990 "name": "BaseBdev3", 00:13:59.990 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:13:59.990 "is_configured": true, 00:13:59.990 "data_offset": 2048, 00:13:59.990 "data_size": 63488 00:13:59.990 } 00:13:59.990 ] 00:13:59.990 }' 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.990 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:00.250 "name": "raid_bdev1", 00:14:00.250 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:00.250 "strip_size_kb": 64, 00:14:00.250 "state": "online", 00:14:00.250 "raid_level": "raid5f", 00:14:00.250 "superblock": true, 00:14:00.250 "num_base_bdevs": 3, 00:14:00.250 "num_base_bdevs_discovered": 3, 00:14:00.250 "num_base_bdevs_operational": 3, 00:14:00.250 "base_bdevs_list": [ 00:14:00.250 { 00:14:00.250 "name": "spare", 00:14:00.250 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 }, 00:14:00.250 { 00:14:00.250 "name": "BaseBdev2", 00:14:00.250 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 }, 00:14:00.250 { 00:14:00.250 "name": "BaseBdev3", 00:14:00.250 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 } 00:14:00.250 ] 00:14:00.250 }' 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.250 "name": "raid_bdev1", 00:14:00.250 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:00.250 "strip_size_kb": 64, 00:14:00.250 "state": "online", 00:14:00.250 "raid_level": "raid5f", 00:14:00.250 "superblock": true, 00:14:00.250 "num_base_bdevs": 3, 00:14:00.250 "num_base_bdevs_discovered": 3, 00:14:00.250 "num_base_bdevs_operational": 3, 00:14:00.250 "base_bdevs_list": [ 00:14:00.250 { 00:14:00.250 "name": "spare", 00:14:00.250 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 }, 00:14:00.250 { 00:14:00.250 "name": "BaseBdev2", 00:14:00.250 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 }, 00:14:00.250 { 00:14:00.250 "name": "BaseBdev3", 00:14:00.250 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:00.250 "is_configured": true, 00:14:00.250 "data_offset": 2048, 00:14:00.250 "data_size": 63488 00:14:00.250 } 00:14:00.250 ] 00:14:00.250 }' 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.250 14:13:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.819 [2024-09-30 14:13:05.227016] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:00.819 [2024-09-30 14:13:05.227047] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:00.819 [2024-09-30 14:13:05.227121] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:00.819 [2024-09-30 14:13:05.227201] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:00.819 [2024-09-30 14:13:05.227221] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:00.819 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:01.079 /dev/nbd0 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:01.079 1+0 records in 00:14:01.079 1+0 records out 00:14:01.079 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000467395 s, 8.8 MB/s 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:01.079 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:01.079 /dev/nbd1 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:01.339 1+0 records in 00:14:01.339 1+0 records out 00:14:01.339 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000328617 s, 12.5 MB/s 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:01.339 14:13:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:01.599 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.859 [2024-09-30 14:13:06.302644] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:01.859 [2024-09-30 14:13:06.302701] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:01.859 [2024-09-30 14:13:06.302736] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:01.859 [2024-09-30 14:13:06.302747] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:01.859 [2024-09-30 14:13:06.304966] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:01.859 [2024-09-30 14:13:06.305004] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:01.859 [2024-09-30 14:13:06.305072] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:01.859 [2024-09-30 14:13:06.305108] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:01.859 [2024-09-30 14:13:06.305212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:01.859 [2024-09-30 14:13:06.305290] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:01.859 spare 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.859 [2024-09-30 14:13:06.405162] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:14:01.859 [2024-09-30 14:13:06.405193] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:01.859 [2024-09-30 14:13:06.405422] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000477d0 00:14:01.859 [2024-09-30 14:13:06.405797] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:14:01.859 [2024-09-30 14:13:06.405820] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:14:01.859 [2024-09-30 14:13:06.405938] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.859 "name": "raid_bdev1", 00:14:01.859 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:01.859 "strip_size_kb": 64, 00:14:01.859 "state": "online", 00:14:01.859 "raid_level": "raid5f", 00:14:01.859 "superblock": true, 00:14:01.859 "num_base_bdevs": 3, 00:14:01.859 "num_base_bdevs_discovered": 3, 00:14:01.859 "num_base_bdevs_operational": 3, 00:14:01.859 "base_bdevs_list": [ 00:14:01.859 { 00:14:01.859 "name": "spare", 00:14:01.859 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:01.859 "is_configured": true, 00:14:01.859 "data_offset": 2048, 00:14:01.859 "data_size": 63488 00:14:01.859 }, 00:14:01.859 { 00:14:01.859 "name": "BaseBdev2", 00:14:01.859 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:01.859 "is_configured": true, 00:14:01.859 "data_offset": 2048, 00:14:01.859 "data_size": 63488 00:14:01.859 }, 00:14:01.859 { 00:14:01.859 "name": "BaseBdev3", 00:14:01.859 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:01.859 "is_configured": true, 00:14:01.859 "data_offset": 2048, 00:14:01.859 "data_size": 63488 00:14:01.859 } 00:14:01.859 ] 00:14:01.859 }' 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.859 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.429 "name": "raid_bdev1", 00:14:02.429 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:02.429 "strip_size_kb": 64, 00:14:02.429 "state": "online", 00:14:02.429 "raid_level": "raid5f", 00:14:02.429 "superblock": true, 00:14:02.429 "num_base_bdevs": 3, 00:14:02.429 "num_base_bdevs_discovered": 3, 00:14:02.429 "num_base_bdevs_operational": 3, 00:14:02.429 "base_bdevs_list": [ 00:14:02.429 { 00:14:02.429 "name": "spare", 00:14:02.429 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev2", 00:14:02.429 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev3", 00:14:02.429 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 } 00:14:02.429 ] 00:14:02.429 }' 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:02.429 14:13:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 [2024-09-30 14:13:07.014053] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.429 "name": "raid_bdev1", 00:14:02.429 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:02.429 "strip_size_kb": 64, 00:14:02.429 "state": "online", 00:14:02.429 "raid_level": "raid5f", 00:14:02.429 "superblock": true, 00:14:02.429 "num_base_bdevs": 3, 00:14:02.429 "num_base_bdevs_discovered": 2, 00:14:02.429 "num_base_bdevs_operational": 2, 00:14:02.429 "base_bdevs_list": [ 00:14:02.429 { 00:14:02.429 "name": null, 00:14:02.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.429 "is_configured": false, 00:14:02.429 "data_offset": 0, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev2", 00:14:02.429 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev3", 00:14:02.429 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 } 00:14:02.429 ] 00:14:02.429 }' 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.429 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.998 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:02.998 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.998 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.998 [2024-09-30 14:13:07.469270] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:02.998 [2024-09-30 14:13:07.469434] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:02.998 [2024-09-30 14:13:07.469457] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:02.998 [2024-09-30 14:13:07.469522] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:02.998 [2024-09-30 14:13:07.472965] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000478a0 00:14:02.998 [2024-09-30 14:13:07.475035] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:02.998 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.998 14:13:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.937 "name": "raid_bdev1", 00:14:03.937 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:03.937 "strip_size_kb": 64, 00:14:03.937 "state": "online", 00:14:03.937 "raid_level": "raid5f", 00:14:03.937 "superblock": true, 00:14:03.937 "num_base_bdevs": 3, 00:14:03.937 "num_base_bdevs_discovered": 3, 00:14:03.937 "num_base_bdevs_operational": 3, 00:14:03.937 "process": { 00:14:03.937 "type": "rebuild", 00:14:03.937 "target": "spare", 00:14:03.937 "progress": { 00:14:03.937 "blocks": 20480, 00:14:03.937 "percent": 16 00:14:03.937 } 00:14:03.937 }, 00:14:03.937 "base_bdevs_list": [ 00:14:03.937 { 00:14:03.937 "name": "spare", 00:14:03.937 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:03.937 "is_configured": true, 00:14:03.937 "data_offset": 2048, 00:14:03.937 "data_size": 63488 00:14:03.937 }, 00:14:03.937 { 00:14:03.937 "name": "BaseBdev2", 00:14:03.937 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:03.937 "is_configured": true, 00:14:03.937 "data_offset": 2048, 00:14:03.937 "data_size": 63488 00:14:03.937 }, 00:14:03.937 { 00:14:03.937 "name": "BaseBdev3", 00:14:03.937 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:03.937 "is_configured": true, 00:14:03.937 "data_offset": 2048, 00:14:03.937 "data_size": 63488 00:14:03.937 } 00:14:03.937 ] 00:14:03.937 }' 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:03.937 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.197 [2024-09-30 14:13:08.641535] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:04.197 [2024-09-30 14:13:08.681586] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:04.197 [2024-09-30 14:13:08.681639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:04.197 [2024-09-30 14:13:08.681653] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:04.197 [2024-09-30 14:13:08.681664] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.197 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.197 "name": "raid_bdev1", 00:14:04.197 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:04.197 "strip_size_kb": 64, 00:14:04.197 "state": "online", 00:14:04.197 "raid_level": "raid5f", 00:14:04.197 "superblock": true, 00:14:04.197 "num_base_bdevs": 3, 00:14:04.197 "num_base_bdevs_discovered": 2, 00:14:04.197 "num_base_bdevs_operational": 2, 00:14:04.197 "base_bdevs_list": [ 00:14:04.197 { 00:14:04.197 "name": null, 00:14:04.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.197 "is_configured": false, 00:14:04.197 "data_offset": 0, 00:14:04.197 "data_size": 63488 00:14:04.197 }, 00:14:04.197 { 00:14:04.197 "name": "BaseBdev2", 00:14:04.197 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:04.197 "is_configured": true, 00:14:04.197 "data_offset": 2048, 00:14:04.197 "data_size": 63488 00:14:04.197 }, 00:14:04.197 { 00:14:04.197 "name": "BaseBdev3", 00:14:04.198 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:04.198 "is_configured": true, 00:14:04.198 "data_offset": 2048, 00:14:04.198 "data_size": 63488 00:14:04.198 } 00:14:04.198 ] 00:14:04.198 }' 00:14:04.198 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.198 14:13:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.457 14:13:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:04.457 14:13:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.457 14:13:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.716 [2024-09-30 14:13:09.117752] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:04.716 [2024-09-30 14:13:09.117838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:04.716 [2024-09-30 14:13:09.117858] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:14:04.716 [2024-09-30 14:13:09.117870] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:04.717 [2024-09-30 14:13:09.118301] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:04.717 [2024-09-30 14:13:09.118328] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:04.717 [2024-09-30 14:13:09.118404] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:04.717 [2024-09-30 14:13:09.118417] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:04.717 [2024-09-30 14:13:09.118427] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:04.717 [2024-09-30 14:13:09.118457] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:04.717 [2024-09-30 14:13:09.121630] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047970 00:14:04.717 [2024-09-30 14:13:09.123679] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:04.717 spare 00:14:04.717 14:13:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.717 14:13:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.655 "name": "raid_bdev1", 00:14:05.655 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:05.655 "strip_size_kb": 64, 00:14:05.655 "state": "online", 00:14:05.655 "raid_level": "raid5f", 00:14:05.655 "superblock": true, 00:14:05.655 "num_base_bdevs": 3, 00:14:05.655 "num_base_bdevs_discovered": 3, 00:14:05.655 "num_base_bdevs_operational": 3, 00:14:05.655 "process": { 00:14:05.655 "type": "rebuild", 00:14:05.655 "target": "spare", 00:14:05.655 "progress": { 00:14:05.655 "blocks": 20480, 00:14:05.655 "percent": 16 00:14:05.655 } 00:14:05.655 }, 00:14:05.655 "base_bdevs_list": [ 00:14:05.655 { 00:14:05.655 "name": "spare", 00:14:05.655 "uuid": "55c5757b-66c2-571a-82b8-08986d016182", 00:14:05.655 "is_configured": true, 00:14:05.655 "data_offset": 2048, 00:14:05.655 "data_size": 63488 00:14:05.655 }, 00:14:05.655 { 00:14:05.655 "name": "BaseBdev2", 00:14:05.655 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:05.655 "is_configured": true, 00:14:05.655 "data_offset": 2048, 00:14:05.655 "data_size": 63488 00:14:05.655 }, 00:14:05.655 { 00:14:05.655 "name": "BaseBdev3", 00:14:05.655 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:05.655 "is_configured": true, 00:14:05.655 "data_offset": 2048, 00:14:05.655 "data_size": 63488 00:14:05.655 } 00:14:05.655 ] 00:14:05.655 }' 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.655 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.655 [2024-09-30 14:13:10.286911] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:05.916 [2024-09-30 14:13:10.330119] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:05.916 [2024-09-30 14:13:10.330170] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:05.916 [2024-09-30 14:13:10.330205] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:05.916 [2024-09-30 14:13:10.330213] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.916 "name": "raid_bdev1", 00:14:05.916 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:05.916 "strip_size_kb": 64, 00:14:05.916 "state": "online", 00:14:05.916 "raid_level": "raid5f", 00:14:05.916 "superblock": true, 00:14:05.916 "num_base_bdevs": 3, 00:14:05.916 "num_base_bdevs_discovered": 2, 00:14:05.916 "num_base_bdevs_operational": 2, 00:14:05.916 "base_bdevs_list": [ 00:14:05.916 { 00:14:05.916 "name": null, 00:14:05.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.916 "is_configured": false, 00:14:05.916 "data_offset": 0, 00:14:05.916 "data_size": 63488 00:14:05.916 }, 00:14:05.916 { 00:14:05.916 "name": "BaseBdev2", 00:14:05.916 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:05.916 "is_configured": true, 00:14:05.916 "data_offset": 2048, 00:14:05.916 "data_size": 63488 00:14:05.916 }, 00:14:05.916 { 00:14:05.916 "name": "BaseBdev3", 00:14:05.916 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:05.916 "is_configured": true, 00:14:05.916 "data_offset": 2048, 00:14:05.916 "data_size": 63488 00:14:05.916 } 00:14:05.916 ] 00:14:05.916 }' 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.916 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.175 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.435 "name": "raid_bdev1", 00:14:06.435 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:06.435 "strip_size_kb": 64, 00:14:06.435 "state": "online", 00:14:06.435 "raid_level": "raid5f", 00:14:06.435 "superblock": true, 00:14:06.435 "num_base_bdevs": 3, 00:14:06.435 "num_base_bdevs_discovered": 2, 00:14:06.435 "num_base_bdevs_operational": 2, 00:14:06.435 "base_bdevs_list": [ 00:14:06.435 { 00:14:06.435 "name": null, 00:14:06.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.435 "is_configured": false, 00:14:06.435 "data_offset": 0, 00:14:06.435 "data_size": 63488 00:14:06.435 }, 00:14:06.435 { 00:14:06.435 "name": "BaseBdev2", 00:14:06.435 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:06.435 "is_configured": true, 00:14:06.435 "data_offset": 2048, 00:14:06.435 "data_size": 63488 00:14:06.435 }, 00:14:06.435 { 00:14:06.435 "name": "BaseBdev3", 00:14:06.435 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:06.435 "is_configured": true, 00:14:06.435 "data_offset": 2048, 00:14:06.435 "data_size": 63488 00:14:06.435 } 00:14:06.435 ] 00:14:06.435 }' 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.435 [2024-09-30 14:13:10.958062] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:06.435 [2024-09-30 14:13:10.958111] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.435 [2024-09-30 14:13:10.958151] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:06.435 [2024-09-30 14:13:10.958160] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.435 [2024-09-30 14:13:10.958575] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.435 [2024-09-30 14:13:10.958603] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:06.435 [2024-09-30 14:13:10.958671] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:06.435 [2024-09-30 14:13:10.958684] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:06.435 [2024-09-30 14:13:10.958694] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:06.435 [2024-09-30 14:13:10.958703] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:06.435 BaseBdev1 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.435 14:13:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.373 14:13:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.373 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.373 "name": "raid_bdev1", 00:14:07.373 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:07.373 "strip_size_kb": 64, 00:14:07.373 "state": "online", 00:14:07.373 "raid_level": "raid5f", 00:14:07.373 "superblock": true, 00:14:07.373 "num_base_bdevs": 3, 00:14:07.373 "num_base_bdevs_discovered": 2, 00:14:07.373 "num_base_bdevs_operational": 2, 00:14:07.373 "base_bdevs_list": [ 00:14:07.373 { 00:14:07.373 "name": null, 00:14:07.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.373 "is_configured": false, 00:14:07.373 "data_offset": 0, 00:14:07.373 "data_size": 63488 00:14:07.373 }, 00:14:07.373 { 00:14:07.373 "name": "BaseBdev2", 00:14:07.373 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:07.373 "is_configured": true, 00:14:07.373 "data_offset": 2048, 00:14:07.373 "data_size": 63488 00:14:07.373 }, 00:14:07.373 { 00:14:07.373 "name": "BaseBdev3", 00:14:07.373 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:07.373 "is_configured": true, 00:14:07.373 "data_offset": 2048, 00:14:07.373 "data_size": 63488 00:14:07.373 } 00:14:07.373 ] 00:14:07.373 }' 00:14:07.373 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.373 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:07.942 "name": "raid_bdev1", 00:14:07.942 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:07.942 "strip_size_kb": 64, 00:14:07.942 "state": "online", 00:14:07.942 "raid_level": "raid5f", 00:14:07.942 "superblock": true, 00:14:07.942 "num_base_bdevs": 3, 00:14:07.942 "num_base_bdevs_discovered": 2, 00:14:07.942 "num_base_bdevs_operational": 2, 00:14:07.942 "base_bdevs_list": [ 00:14:07.942 { 00:14:07.942 "name": null, 00:14:07.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.942 "is_configured": false, 00:14:07.942 "data_offset": 0, 00:14:07.942 "data_size": 63488 00:14:07.942 }, 00:14:07.942 { 00:14:07.942 "name": "BaseBdev2", 00:14:07.942 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:07.942 "is_configured": true, 00:14:07.942 "data_offset": 2048, 00:14:07.942 "data_size": 63488 00:14:07.942 }, 00:14:07.942 { 00:14:07.942 "name": "BaseBdev3", 00:14:07.942 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:07.942 "is_configured": true, 00:14:07.942 "data_offset": 2048, 00:14:07.942 "data_size": 63488 00:14:07.942 } 00:14:07.942 ] 00:14:07.942 }' 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.942 [2024-09-30 14:13:12.547386] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:07.942 [2024-09-30 14:13:12.547556] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:07.942 [2024-09-30 14:13:12.547578] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:07.942 request: 00:14:07.942 { 00:14:07.942 "base_bdev": "BaseBdev1", 00:14:07.942 "raid_bdev": "raid_bdev1", 00:14:07.942 "method": "bdev_raid_add_base_bdev", 00:14:07.942 "req_id": 1 00:14:07.942 } 00:14:07.942 Got JSON-RPC error response 00:14:07.942 response: 00:14:07.942 { 00:14:07.942 "code": -22, 00:14:07.942 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:07.942 } 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:07.942 14:13:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.323 "name": "raid_bdev1", 00:14:09.323 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:09.323 "strip_size_kb": 64, 00:14:09.323 "state": "online", 00:14:09.323 "raid_level": "raid5f", 00:14:09.323 "superblock": true, 00:14:09.323 "num_base_bdevs": 3, 00:14:09.323 "num_base_bdevs_discovered": 2, 00:14:09.323 "num_base_bdevs_operational": 2, 00:14:09.323 "base_bdevs_list": [ 00:14:09.323 { 00:14:09.323 "name": null, 00:14:09.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.323 "is_configured": false, 00:14:09.323 "data_offset": 0, 00:14:09.323 "data_size": 63488 00:14:09.323 }, 00:14:09.323 { 00:14:09.323 "name": "BaseBdev2", 00:14:09.323 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:09.323 "is_configured": true, 00:14:09.323 "data_offset": 2048, 00:14:09.323 "data_size": 63488 00:14:09.323 }, 00:14:09.323 { 00:14:09.323 "name": "BaseBdev3", 00:14:09.323 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:09.323 "is_configured": true, 00:14:09.323 "data_offset": 2048, 00:14:09.323 "data_size": 63488 00:14:09.323 } 00:14:09.323 ] 00:14:09.323 }' 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.323 14:13:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.584 "name": "raid_bdev1", 00:14:09.584 "uuid": "1cf558bb-6947-45be-aba7-96acf69a0fd2", 00:14:09.584 "strip_size_kb": 64, 00:14:09.584 "state": "online", 00:14:09.584 "raid_level": "raid5f", 00:14:09.584 "superblock": true, 00:14:09.584 "num_base_bdevs": 3, 00:14:09.584 "num_base_bdevs_discovered": 2, 00:14:09.584 "num_base_bdevs_operational": 2, 00:14:09.584 "base_bdevs_list": [ 00:14:09.584 { 00:14:09.584 "name": null, 00:14:09.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.584 "is_configured": false, 00:14:09.584 "data_offset": 0, 00:14:09.584 "data_size": 63488 00:14:09.584 }, 00:14:09.584 { 00:14:09.584 "name": "BaseBdev2", 00:14:09.584 "uuid": "2177ba49-8a74-5e66-bcc6-79959a11491b", 00:14:09.584 "is_configured": true, 00:14:09.584 "data_offset": 2048, 00:14:09.584 "data_size": 63488 00:14:09.584 }, 00:14:09.584 { 00:14:09.584 "name": "BaseBdev3", 00:14:09.584 "uuid": "fd330a09-1087-5b27-957f-f5c263c7785c", 00:14:09.584 "is_configured": true, 00:14:09.584 "data_offset": 2048, 00:14:09.584 "data_size": 63488 00:14:09.584 } 00:14:09.584 ] 00:14:09.584 }' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 93533 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 93533 ']' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 93533 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 93533 00:14:09.584 killing process with pid 93533 00:14:09.584 Received shutdown signal, test time was about 60.000000 seconds 00:14:09.584 00:14:09.584 Latency(us) 00:14:09.584 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:09.584 =================================================================================================================== 00:14:09.584 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 93533' 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 93533 00:14:09.584 [2024-09-30 14:13:14.218729] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:09.584 [2024-09-30 14:13:14.218856] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:09.584 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 93533 00:14:09.584 [2024-09-30 14:13:14.218915] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:09.584 [2024-09-30 14:13:14.218927] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:14:09.845 [2024-09-30 14:13:14.259898] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:09.845 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:09.845 00:14:09.845 real 0m21.591s 00:14:09.845 user 0m28.096s 00:14:09.845 sys 0m2.807s 00:14:09.845 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:09.845 14:13:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.845 ************************************ 00:14:09.845 END TEST raid5f_rebuild_test_sb 00:14:09.845 ************************************ 00:14:10.105 14:13:14 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:10.105 14:13:14 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:10.105 14:13:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:10.105 14:13:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:10.105 14:13:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:10.105 ************************************ 00:14:10.105 START TEST raid5f_state_function_test 00:14:10.105 ************************************ 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 false 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:10.105 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=94264 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 94264' 00:14:10.106 Process raid pid: 94264 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 94264 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 94264 ']' 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:10.106 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:10.106 14:13:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.106 [2024-09-30 14:13:14.668465] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:14:10.106 [2024-09-30 14:13:14.668603] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:10.366 [2024-09-30 14:13:14.801776] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:14:10.366 [2024-09-30 14:13:14.830148] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:10.366 [2024-09-30 14:13:14.875755] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:10.366 [2024-09-30 14:13:14.917600] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:10.366 [2024-09-30 14:13:14.917642] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.936 [2024-09-30 14:13:15.490748] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:10.936 [2024-09-30 14:13:15.490813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:10.936 [2024-09-30 14:13:15.490825] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:10.936 [2024-09-30 14:13:15.490832] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:10.936 [2024-09-30 14:13:15.490842] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:10.936 [2024-09-30 14:13:15.490849] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:10.936 [2024-09-30 14:13:15.490856] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:10.936 [2024-09-30 14:13:15.490863] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.936 "name": "Existed_Raid", 00:14:10.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.936 "strip_size_kb": 64, 00:14:10.936 "state": "configuring", 00:14:10.936 "raid_level": "raid5f", 00:14:10.936 "superblock": false, 00:14:10.936 "num_base_bdevs": 4, 00:14:10.936 "num_base_bdevs_discovered": 0, 00:14:10.936 "num_base_bdevs_operational": 4, 00:14:10.936 "base_bdevs_list": [ 00:14:10.936 { 00:14:10.936 "name": "BaseBdev1", 00:14:10.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.936 "is_configured": false, 00:14:10.936 "data_offset": 0, 00:14:10.936 "data_size": 0 00:14:10.936 }, 00:14:10.936 { 00:14:10.936 "name": "BaseBdev2", 00:14:10.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.936 "is_configured": false, 00:14:10.936 "data_offset": 0, 00:14:10.936 "data_size": 0 00:14:10.936 }, 00:14:10.936 { 00:14:10.936 "name": "BaseBdev3", 00:14:10.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.936 "is_configured": false, 00:14:10.936 "data_offset": 0, 00:14:10.936 "data_size": 0 00:14:10.936 }, 00:14:10.936 { 00:14:10.936 "name": "BaseBdev4", 00:14:10.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.936 "is_configured": false, 00:14:10.936 "data_offset": 0, 00:14:10.936 "data_size": 0 00:14:10.936 } 00:14:10.936 ] 00:14:10.936 }' 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.936 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 [2024-09-30 14:13:15.965920] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:11.507 [2024-09-30 14:13:15.965961] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 [2024-09-30 14:13:15.977899] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:11.507 [2024-09-30 14:13:15.977936] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:11.507 [2024-09-30 14:13:15.977962] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:11.507 [2024-09-30 14:13:15.977969] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:11.507 [2024-09-30 14:13:15.977977] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:11.507 [2024-09-30 14:13:15.977984] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:11.507 [2024-09-30 14:13:15.977991] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:11.507 [2024-09-30 14:13:15.977997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 [2024-09-30 14:13:15.998566] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:11.507 BaseBdev1 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:11.507 14:13:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 [ 00:14:11.507 { 00:14:11.507 "name": "BaseBdev1", 00:14:11.507 "aliases": [ 00:14:11.507 "f0f3cc26-f893-476d-821b-bd01f79f89e9" 00:14:11.507 ], 00:14:11.507 "product_name": "Malloc disk", 00:14:11.507 "block_size": 512, 00:14:11.507 "num_blocks": 65536, 00:14:11.507 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:11.507 "assigned_rate_limits": { 00:14:11.507 "rw_ios_per_sec": 0, 00:14:11.507 "rw_mbytes_per_sec": 0, 00:14:11.507 "r_mbytes_per_sec": 0, 00:14:11.507 "w_mbytes_per_sec": 0 00:14:11.507 }, 00:14:11.507 "claimed": true, 00:14:11.507 "claim_type": "exclusive_write", 00:14:11.507 "zoned": false, 00:14:11.507 "supported_io_types": { 00:14:11.507 "read": true, 00:14:11.507 "write": true, 00:14:11.507 "unmap": true, 00:14:11.507 "flush": true, 00:14:11.507 "reset": true, 00:14:11.507 "nvme_admin": false, 00:14:11.507 "nvme_io": false, 00:14:11.507 "nvme_io_md": false, 00:14:11.507 "write_zeroes": true, 00:14:11.507 "zcopy": true, 00:14:11.507 "get_zone_info": false, 00:14:11.507 "zone_management": false, 00:14:11.507 "zone_append": false, 00:14:11.507 "compare": false, 00:14:11.507 "compare_and_write": false, 00:14:11.507 "abort": true, 00:14:11.507 "seek_hole": false, 00:14:11.507 "seek_data": false, 00:14:11.507 "copy": true, 00:14:11.507 "nvme_iov_md": false 00:14:11.507 }, 00:14:11.507 "memory_domains": [ 00:14:11.507 { 00:14:11.507 "dma_device_id": "system", 00:14:11.507 "dma_device_type": 1 00:14:11.507 }, 00:14:11.507 { 00:14:11.507 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:11.507 "dma_device_type": 2 00:14:11.507 } 00:14:11.507 ], 00:14:11.507 "driver_specific": {} 00:14:11.507 } 00:14:11.507 ] 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.507 "name": "Existed_Raid", 00:14:11.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.507 "strip_size_kb": 64, 00:14:11.507 "state": "configuring", 00:14:11.507 "raid_level": "raid5f", 00:14:11.507 "superblock": false, 00:14:11.507 "num_base_bdevs": 4, 00:14:11.507 "num_base_bdevs_discovered": 1, 00:14:11.507 "num_base_bdevs_operational": 4, 00:14:11.507 "base_bdevs_list": [ 00:14:11.507 { 00:14:11.507 "name": "BaseBdev1", 00:14:11.507 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:11.507 "is_configured": true, 00:14:11.507 "data_offset": 0, 00:14:11.507 "data_size": 65536 00:14:11.507 }, 00:14:11.507 { 00:14:11.507 "name": "BaseBdev2", 00:14:11.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.507 "is_configured": false, 00:14:11.507 "data_offset": 0, 00:14:11.507 "data_size": 0 00:14:11.507 }, 00:14:11.507 { 00:14:11.507 "name": "BaseBdev3", 00:14:11.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.507 "is_configured": false, 00:14:11.507 "data_offset": 0, 00:14:11.507 "data_size": 0 00:14:11.507 }, 00:14:11.507 { 00:14:11.507 "name": "BaseBdev4", 00:14:11.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.507 "is_configured": false, 00:14:11.507 "data_offset": 0, 00:14:11.507 "data_size": 0 00:14:11.507 } 00:14:11.507 ] 00:14:11.507 }' 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.507 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.077 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:12.077 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.077 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.078 [2024-09-30 14:13:16.453781] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:12.078 [2024-09-30 14:13:16.453836] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.078 [2024-09-30 14:13:16.465781] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:12.078 [2024-09-30 14:13:16.467533] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:12.078 [2024-09-30 14:13:16.467566] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:12.078 [2024-09-30 14:13:16.467592] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:12.078 [2024-09-30 14:13:16.467600] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:12.078 [2024-09-30 14:13:16.467607] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:12.078 [2024-09-30 14:13:16.467614] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.078 "name": "Existed_Raid", 00:14:12.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.078 "strip_size_kb": 64, 00:14:12.078 "state": "configuring", 00:14:12.078 "raid_level": "raid5f", 00:14:12.078 "superblock": false, 00:14:12.078 "num_base_bdevs": 4, 00:14:12.078 "num_base_bdevs_discovered": 1, 00:14:12.078 "num_base_bdevs_operational": 4, 00:14:12.078 "base_bdevs_list": [ 00:14:12.078 { 00:14:12.078 "name": "BaseBdev1", 00:14:12.078 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:12.078 "is_configured": true, 00:14:12.078 "data_offset": 0, 00:14:12.078 "data_size": 65536 00:14:12.078 }, 00:14:12.078 { 00:14:12.078 "name": "BaseBdev2", 00:14:12.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.078 "is_configured": false, 00:14:12.078 "data_offset": 0, 00:14:12.078 "data_size": 0 00:14:12.078 }, 00:14:12.078 { 00:14:12.078 "name": "BaseBdev3", 00:14:12.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.078 "is_configured": false, 00:14:12.078 "data_offset": 0, 00:14:12.078 "data_size": 0 00:14:12.078 }, 00:14:12.078 { 00:14:12.078 "name": "BaseBdev4", 00:14:12.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.078 "is_configured": false, 00:14:12.078 "data_offset": 0, 00:14:12.078 "data_size": 0 00:14:12.078 } 00:14:12.078 ] 00:14:12.078 }' 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.078 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.338 [2024-09-30 14:13:16.919035] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:12.338 BaseBdev2 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.338 [ 00:14:12.338 { 00:14:12.338 "name": "BaseBdev2", 00:14:12.338 "aliases": [ 00:14:12.338 "637daf57-37db-4cb7-815b-7fbb9902b9f6" 00:14:12.338 ], 00:14:12.338 "product_name": "Malloc disk", 00:14:12.338 "block_size": 512, 00:14:12.338 "num_blocks": 65536, 00:14:12.338 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:12.338 "assigned_rate_limits": { 00:14:12.338 "rw_ios_per_sec": 0, 00:14:12.338 "rw_mbytes_per_sec": 0, 00:14:12.338 "r_mbytes_per_sec": 0, 00:14:12.338 "w_mbytes_per_sec": 0 00:14:12.338 }, 00:14:12.338 "claimed": true, 00:14:12.338 "claim_type": "exclusive_write", 00:14:12.338 "zoned": false, 00:14:12.338 "supported_io_types": { 00:14:12.338 "read": true, 00:14:12.338 "write": true, 00:14:12.338 "unmap": true, 00:14:12.338 "flush": true, 00:14:12.338 "reset": true, 00:14:12.338 "nvme_admin": false, 00:14:12.338 "nvme_io": false, 00:14:12.338 "nvme_io_md": false, 00:14:12.338 "write_zeroes": true, 00:14:12.338 "zcopy": true, 00:14:12.338 "get_zone_info": false, 00:14:12.338 "zone_management": false, 00:14:12.338 "zone_append": false, 00:14:12.338 "compare": false, 00:14:12.338 "compare_and_write": false, 00:14:12.338 "abort": true, 00:14:12.338 "seek_hole": false, 00:14:12.338 "seek_data": false, 00:14:12.338 "copy": true, 00:14:12.338 "nvme_iov_md": false 00:14:12.338 }, 00:14:12.338 "memory_domains": [ 00:14:12.338 { 00:14:12.338 "dma_device_id": "system", 00:14:12.338 "dma_device_type": 1 00:14:12.338 }, 00:14:12.338 { 00:14:12.338 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.338 "dma_device_type": 2 00:14:12.338 } 00:14:12.338 ], 00:14:12.338 "driver_specific": {} 00:14:12.338 } 00:14:12.338 ] 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.338 14:13:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.598 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.598 "name": "Existed_Raid", 00:14:12.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.598 "strip_size_kb": 64, 00:14:12.598 "state": "configuring", 00:14:12.598 "raid_level": "raid5f", 00:14:12.598 "superblock": false, 00:14:12.598 "num_base_bdevs": 4, 00:14:12.598 "num_base_bdevs_discovered": 2, 00:14:12.598 "num_base_bdevs_operational": 4, 00:14:12.598 "base_bdevs_list": [ 00:14:12.598 { 00:14:12.598 "name": "BaseBdev1", 00:14:12.598 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:12.598 "is_configured": true, 00:14:12.598 "data_offset": 0, 00:14:12.598 "data_size": 65536 00:14:12.598 }, 00:14:12.598 { 00:14:12.598 "name": "BaseBdev2", 00:14:12.598 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:12.598 "is_configured": true, 00:14:12.598 "data_offset": 0, 00:14:12.598 "data_size": 65536 00:14:12.598 }, 00:14:12.598 { 00:14:12.598 "name": "BaseBdev3", 00:14:12.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.598 "is_configured": false, 00:14:12.598 "data_offset": 0, 00:14:12.598 "data_size": 0 00:14:12.598 }, 00:14:12.598 { 00:14:12.598 "name": "BaseBdev4", 00:14:12.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.598 "is_configured": false, 00:14:12.598 "data_offset": 0, 00:14:12.598 "data_size": 0 00:14:12.599 } 00:14:12.599 ] 00:14:12.599 }' 00:14:12.599 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.599 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.859 [2024-09-30 14:13:17.381238] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:12.859 BaseBdev3 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.859 [ 00:14:12.859 { 00:14:12.859 "name": "BaseBdev3", 00:14:12.859 "aliases": [ 00:14:12.859 "db63baf0-c6da-4bb1-99cf-34ea728aaac1" 00:14:12.859 ], 00:14:12.859 "product_name": "Malloc disk", 00:14:12.859 "block_size": 512, 00:14:12.859 "num_blocks": 65536, 00:14:12.859 "uuid": "db63baf0-c6da-4bb1-99cf-34ea728aaac1", 00:14:12.859 "assigned_rate_limits": { 00:14:12.859 "rw_ios_per_sec": 0, 00:14:12.859 "rw_mbytes_per_sec": 0, 00:14:12.859 "r_mbytes_per_sec": 0, 00:14:12.859 "w_mbytes_per_sec": 0 00:14:12.859 }, 00:14:12.859 "claimed": true, 00:14:12.859 "claim_type": "exclusive_write", 00:14:12.859 "zoned": false, 00:14:12.859 "supported_io_types": { 00:14:12.859 "read": true, 00:14:12.859 "write": true, 00:14:12.859 "unmap": true, 00:14:12.859 "flush": true, 00:14:12.859 "reset": true, 00:14:12.859 "nvme_admin": false, 00:14:12.859 "nvme_io": false, 00:14:12.859 "nvme_io_md": false, 00:14:12.859 "write_zeroes": true, 00:14:12.859 "zcopy": true, 00:14:12.859 "get_zone_info": false, 00:14:12.859 "zone_management": false, 00:14:12.859 "zone_append": false, 00:14:12.859 "compare": false, 00:14:12.859 "compare_and_write": false, 00:14:12.859 "abort": true, 00:14:12.859 "seek_hole": false, 00:14:12.859 "seek_data": false, 00:14:12.859 "copy": true, 00:14:12.859 "nvme_iov_md": false 00:14:12.859 }, 00:14:12.859 "memory_domains": [ 00:14:12.859 { 00:14:12.859 "dma_device_id": "system", 00:14:12.859 "dma_device_type": 1 00:14:12.859 }, 00:14:12.859 { 00:14:12.859 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.859 "dma_device_type": 2 00:14:12.859 } 00:14:12.859 ], 00:14:12.859 "driver_specific": {} 00:14:12.859 } 00:14:12.859 ] 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.859 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.860 "name": "Existed_Raid", 00:14:12.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.860 "strip_size_kb": 64, 00:14:12.860 "state": "configuring", 00:14:12.860 "raid_level": "raid5f", 00:14:12.860 "superblock": false, 00:14:12.860 "num_base_bdevs": 4, 00:14:12.860 "num_base_bdevs_discovered": 3, 00:14:12.860 "num_base_bdevs_operational": 4, 00:14:12.860 "base_bdevs_list": [ 00:14:12.860 { 00:14:12.860 "name": "BaseBdev1", 00:14:12.860 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:12.860 "is_configured": true, 00:14:12.860 "data_offset": 0, 00:14:12.860 "data_size": 65536 00:14:12.860 }, 00:14:12.860 { 00:14:12.860 "name": "BaseBdev2", 00:14:12.860 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:12.860 "is_configured": true, 00:14:12.860 "data_offset": 0, 00:14:12.860 "data_size": 65536 00:14:12.860 }, 00:14:12.860 { 00:14:12.860 "name": "BaseBdev3", 00:14:12.860 "uuid": "db63baf0-c6da-4bb1-99cf-34ea728aaac1", 00:14:12.860 "is_configured": true, 00:14:12.860 "data_offset": 0, 00:14:12.860 "data_size": 65536 00:14:12.860 }, 00:14:12.860 { 00:14:12.860 "name": "BaseBdev4", 00:14:12.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.860 "is_configured": false, 00:14:12.860 "data_offset": 0, 00:14:12.860 "data_size": 0 00:14:12.860 } 00:14:12.860 ] 00:14:12.860 }' 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.860 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.430 [2024-09-30 14:13:17.879435] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:13.430 [2024-09-30 14:13:17.879500] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:14:13.430 [2024-09-30 14:13:17.879511] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:13.430 [2024-09-30 14:13:17.879774] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:13.430 [2024-09-30 14:13:17.880217] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:14:13.430 [2024-09-30 14:13:17.880255] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:14:13.430 [2024-09-30 14:13:17.880461] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.430 BaseBdev4 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.430 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.431 [ 00:14:13.431 { 00:14:13.431 "name": "BaseBdev4", 00:14:13.431 "aliases": [ 00:14:13.431 "c03e2916-6264-4541-b934-388f22bd9aa1" 00:14:13.431 ], 00:14:13.431 "product_name": "Malloc disk", 00:14:13.431 "block_size": 512, 00:14:13.431 "num_blocks": 65536, 00:14:13.431 "uuid": "c03e2916-6264-4541-b934-388f22bd9aa1", 00:14:13.431 "assigned_rate_limits": { 00:14:13.431 "rw_ios_per_sec": 0, 00:14:13.431 "rw_mbytes_per_sec": 0, 00:14:13.431 "r_mbytes_per_sec": 0, 00:14:13.431 "w_mbytes_per_sec": 0 00:14:13.431 }, 00:14:13.431 "claimed": true, 00:14:13.431 "claim_type": "exclusive_write", 00:14:13.431 "zoned": false, 00:14:13.431 "supported_io_types": { 00:14:13.431 "read": true, 00:14:13.431 "write": true, 00:14:13.431 "unmap": true, 00:14:13.431 "flush": true, 00:14:13.431 "reset": true, 00:14:13.431 "nvme_admin": false, 00:14:13.431 "nvme_io": false, 00:14:13.431 "nvme_io_md": false, 00:14:13.431 "write_zeroes": true, 00:14:13.431 "zcopy": true, 00:14:13.431 "get_zone_info": false, 00:14:13.431 "zone_management": false, 00:14:13.431 "zone_append": false, 00:14:13.431 "compare": false, 00:14:13.431 "compare_and_write": false, 00:14:13.431 "abort": true, 00:14:13.431 "seek_hole": false, 00:14:13.431 "seek_data": false, 00:14:13.431 "copy": true, 00:14:13.431 "nvme_iov_md": false 00:14:13.431 }, 00:14:13.431 "memory_domains": [ 00:14:13.431 { 00:14:13.431 "dma_device_id": "system", 00:14:13.431 "dma_device_type": 1 00:14:13.431 }, 00:14:13.431 { 00:14:13.431 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:13.431 "dma_device_type": 2 00:14:13.431 } 00:14:13.431 ], 00:14:13.431 "driver_specific": {} 00:14:13.431 } 00:14:13.431 ] 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.431 "name": "Existed_Raid", 00:14:13.431 "uuid": "8793a1b6-ee7d-4ef2-9ba9-f3798c36242f", 00:14:13.431 "strip_size_kb": 64, 00:14:13.431 "state": "online", 00:14:13.431 "raid_level": "raid5f", 00:14:13.431 "superblock": false, 00:14:13.431 "num_base_bdevs": 4, 00:14:13.431 "num_base_bdevs_discovered": 4, 00:14:13.431 "num_base_bdevs_operational": 4, 00:14:13.431 "base_bdevs_list": [ 00:14:13.431 { 00:14:13.431 "name": "BaseBdev1", 00:14:13.431 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:13.431 "is_configured": true, 00:14:13.431 "data_offset": 0, 00:14:13.431 "data_size": 65536 00:14:13.431 }, 00:14:13.431 { 00:14:13.431 "name": "BaseBdev2", 00:14:13.431 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:13.431 "is_configured": true, 00:14:13.431 "data_offset": 0, 00:14:13.431 "data_size": 65536 00:14:13.431 }, 00:14:13.431 { 00:14:13.431 "name": "BaseBdev3", 00:14:13.431 "uuid": "db63baf0-c6da-4bb1-99cf-34ea728aaac1", 00:14:13.431 "is_configured": true, 00:14:13.431 "data_offset": 0, 00:14:13.431 "data_size": 65536 00:14:13.431 }, 00:14:13.431 { 00:14:13.431 "name": "BaseBdev4", 00:14:13.431 "uuid": "c03e2916-6264-4541-b934-388f22bd9aa1", 00:14:13.431 "is_configured": true, 00:14:13.431 "data_offset": 0, 00:14:13.431 "data_size": 65536 00:14:13.431 } 00:14:13.431 ] 00:14:13.431 }' 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.431 14:13:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.999 [2024-09-30 14:13:18.374949] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.999 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:13.999 "name": "Existed_Raid", 00:14:13.999 "aliases": [ 00:14:13.999 "8793a1b6-ee7d-4ef2-9ba9-f3798c36242f" 00:14:13.999 ], 00:14:13.999 "product_name": "Raid Volume", 00:14:13.999 "block_size": 512, 00:14:13.999 "num_blocks": 196608, 00:14:13.999 "uuid": "8793a1b6-ee7d-4ef2-9ba9-f3798c36242f", 00:14:13.999 "assigned_rate_limits": { 00:14:13.999 "rw_ios_per_sec": 0, 00:14:14.000 "rw_mbytes_per_sec": 0, 00:14:14.000 "r_mbytes_per_sec": 0, 00:14:14.000 "w_mbytes_per_sec": 0 00:14:14.000 }, 00:14:14.000 "claimed": false, 00:14:14.000 "zoned": false, 00:14:14.000 "supported_io_types": { 00:14:14.000 "read": true, 00:14:14.000 "write": true, 00:14:14.000 "unmap": false, 00:14:14.000 "flush": false, 00:14:14.000 "reset": true, 00:14:14.000 "nvme_admin": false, 00:14:14.000 "nvme_io": false, 00:14:14.000 "nvme_io_md": false, 00:14:14.000 "write_zeroes": true, 00:14:14.000 "zcopy": false, 00:14:14.000 "get_zone_info": false, 00:14:14.000 "zone_management": false, 00:14:14.000 "zone_append": false, 00:14:14.000 "compare": false, 00:14:14.000 "compare_and_write": false, 00:14:14.000 "abort": false, 00:14:14.000 "seek_hole": false, 00:14:14.000 "seek_data": false, 00:14:14.000 "copy": false, 00:14:14.000 "nvme_iov_md": false 00:14:14.000 }, 00:14:14.000 "driver_specific": { 00:14:14.000 "raid": { 00:14:14.000 "uuid": "8793a1b6-ee7d-4ef2-9ba9-f3798c36242f", 00:14:14.000 "strip_size_kb": 64, 00:14:14.000 "state": "online", 00:14:14.000 "raid_level": "raid5f", 00:14:14.000 "superblock": false, 00:14:14.000 "num_base_bdevs": 4, 00:14:14.000 "num_base_bdevs_discovered": 4, 00:14:14.000 "num_base_bdevs_operational": 4, 00:14:14.000 "base_bdevs_list": [ 00:14:14.000 { 00:14:14.000 "name": "BaseBdev1", 00:14:14.000 "uuid": "f0f3cc26-f893-476d-821b-bd01f79f89e9", 00:14:14.000 "is_configured": true, 00:14:14.000 "data_offset": 0, 00:14:14.000 "data_size": 65536 00:14:14.000 }, 00:14:14.000 { 00:14:14.000 "name": "BaseBdev2", 00:14:14.000 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:14.000 "is_configured": true, 00:14:14.000 "data_offset": 0, 00:14:14.000 "data_size": 65536 00:14:14.000 }, 00:14:14.000 { 00:14:14.000 "name": "BaseBdev3", 00:14:14.000 "uuid": "db63baf0-c6da-4bb1-99cf-34ea728aaac1", 00:14:14.000 "is_configured": true, 00:14:14.000 "data_offset": 0, 00:14:14.000 "data_size": 65536 00:14:14.000 }, 00:14:14.000 { 00:14:14.000 "name": "BaseBdev4", 00:14:14.000 "uuid": "c03e2916-6264-4541-b934-388f22bd9aa1", 00:14:14.000 "is_configured": true, 00:14:14.000 "data_offset": 0, 00:14:14.000 "data_size": 65536 00:14:14.000 } 00:14:14.000 ] 00:14:14.000 } 00:14:14.000 } 00:14:14.000 }' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:14.000 BaseBdev2 00:14:14.000 BaseBdev3 00:14:14.000 BaseBdev4' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.000 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.260 [2024-09-30 14:13:18.670364] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.260 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.260 "name": "Existed_Raid", 00:14:14.260 "uuid": "8793a1b6-ee7d-4ef2-9ba9-f3798c36242f", 00:14:14.260 "strip_size_kb": 64, 00:14:14.260 "state": "online", 00:14:14.260 "raid_level": "raid5f", 00:14:14.260 "superblock": false, 00:14:14.260 "num_base_bdevs": 4, 00:14:14.261 "num_base_bdevs_discovered": 3, 00:14:14.261 "num_base_bdevs_operational": 3, 00:14:14.261 "base_bdevs_list": [ 00:14:14.261 { 00:14:14.261 "name": null, 00:14:14.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.261 "is_configured": false, 00:14:14.261 "data_offset": 0, 00:14:14.261 "data_size": 65536 00:14:14.261 }, 00:14:14.261 { 00:14:14.261 "name": "BaseBdev2", 00:14:14.261 "uuid": "637daf57-37db-4cb7-815b-7fbb9902b9f6", 00:14:14.261 "is_configured": true, 00:14:14.261 "data_offset": 0, 00:14:14.261 "data_size": 65536 00:14:14.261 }, 00:14:14.261 { 00:14:14.261 "name": "BaseBdev3", 00:14:14.261 "uuid": "db63baf0-c6da-4bb1-99cf-34ea728aaac1", 00:14:14.261 "is_configured": true, 00:14:14.261 "data_offset": 0, 00:14:14.261 "data_size": 65536 00:14:14.261 }, 00:14:14.261 { 00:14:14.261 "name": "BaseBdev4", 00:14:14.261 "uuid": "c03e2916-6264-4541-b934-388f22bd9aa1", 00:14:14.261 "is_configured": true, 00:14:14.261 "data_offset": 0, 00:14:14.261 "data_size": 65536 00:14:14.261 } 00:14:14.261 ] 00:14:14.261 }' 00:14:14.261 14:13:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.261 14:13:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.520 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.779 [2024-09-30 14:13:19.176909] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:14.779 [2024-09-30 14:13:19.177005] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:14.780 [2024-09-30 14:13:19.187962] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 [2024-09-30 14:13:19.247929] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 [2024-09-30 14:13:19.314628] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:14.780 [2024-09-30 14:13:19.314689] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 BaseBdev2 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.780 [ 00:14:14.780 { 00:14:14.780 "name": "BaseBdev2", 00:14:14.780 "aliases": [ 00:14:14.780 "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf" 00:14:14.780 ], 00:14:14.780 "product_name": "Malloc disk", 00:14:14.780 "block_size": 512, 00:14:14.780 "num_blocks": 65536, 00:14:14.780 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:14.780 "assigned_rate_limits": { 00:14:14.780 "rw_ios_per_sec": 0, 00:14:14.780 "rw_mbytes_per_sec": 0, 00:14:14.780 "r_mbytes_per_sec": 0, 00:14:14.780 "w_mbytes_per_sec": 0 00:14:14.780 }, 00:14:14.780 "claimed": false, 00:14:14.780 "zoned": false, 00:14:14.780 "supported_io_types": { 00:14:14.780 "read": true, 00:14:14.780 "write": true, 00:14:14.780 "unmap": true, 00:14:14.780 "flush": true, 00:14:14.780 "reset": true, 00:14:14.780 "nvme_admin": false, 00:14:14.780 "nvme_io": false, 00:14:14.780 "nvme_io_md": false, 00:14:14.780 "write_zeroes": true, 00:14:14.780 "zcopy": true, 00:14:14.780 "get_zone_info": false, 00:14:14.780 "zone_management": false, 00:14:14.780 "zone_append": false, 00:14:14.780 "compare": false, 00:14:14.780 "compare_and_write": false, 00:14:14.780 "abort": true, 00:14:14.780 "seek_hole": false, 00:14:14.780 "seek_data": false, 00:14:14.780 "copy": true, 00:14:14.780 "nvme_iov_md": false 00:14:14.780 }, 00:14:14.780 "memory_domains": [ 00:14:14.780 { 00:14:14.780 "dma_device_id": "system", 00:14:14.780 "dma_device_type": 1 00:14:14.780 }, 00:14:14.780 { 00:14:14.780 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:14.780 "dma_device_type": 2 00:14:14.780 } 00:14:14.780 ], 00:14:14.780 "driver_specific": {} 00:14:14.780 } 00:14:14.780 ] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.780 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.040 BaseBdev3 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.040 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.040 [ 00:14:15.040 { 00:14:15.040 "name": "BaseBdev3", 00:14:15.040 "aliases": [ 00:14:15.040 "17db7850-ce13-4d99-9734-673e39eb206f" 00:14:15.040 ], 00:14:15.040 "product_name": "Malloc disk", 00:14:15.040 "block_size": 512, 00:14:15.040 "num_blocks": 65536, 00:14:15.040 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:15.040 "assigned_rate_limits": { 00:14:15.041 "rw_ios_per_sec": 0, 00:14:15.041 "rw_mbytes_per_sec": 0, 00:14:15.041 "r_mbytes_per_sec": 0, 00:14:15.041 "w_mbytes_per_sec": 0 00:14:15.041 }, 00:14:15.041 "claimed": false, 00:14:15.041 "zoned": false, 00:14:15.041 "supported_io_types": { 00:14:15.041 "read": true, 00:14:15.041 "write": true, 00:14:15.041 "unmap": true, 00:14:15.041 "flush": true, 00:14:15.041 "reset": true, 00:14:15.041 "nvme_admin": false, 00:14:15.041 "nvme_io": false, 00:14:15.041 "nvme_io_md": false, 00:14:15.041 "write_zeroes": true, 00:14:15.041 "zcopy": true, 00:14:15.041 "get_zone_info": false, 00:14:15.041 "zone_management": false, 00:14:15.041 "zone_append": false, 00:14:15.041 "compare": false, 00:14:15.041 "compare_and_write": false, 00:14:15.041 "abort": true, 00:14:15.041 "seek_hole": false, 00:14:15.041 "seek_data": false, 00:14:15.041 "copy": true, 00:14:15.041 "nvme_iov_md": false 00:14:15.041 }, 00:14:15.041 "memory_domains": [ 00:14:15.041 { 00:14:15.041 "dma_device_id": "system", 00:14:15.041 "dma_device_type": 1 00:14:15.041 }, 00:14:15.041 { 00:14:15.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:15.041 "dma_device_type": 2 00:14:15.041 } 00:14:15.041 ], 00:14:15.041 "driver_specific": {} 00:14:15.041 } 00:14:15.041 ] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.041 BaseBdev4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.041 [ 00:14:15.041 { 00:14:15.041 "name": "BaseBdev4", 00:14:15.041 "aliases": [ 00:14:15.041 "8b3352b1-8d94-4cab-933d-b5602bc93b96" 00:14:15.041 ], 00:14:15.041 "product_name": "Malloc disk", 00:14:15.041 "block_size": 512, 00:14:15.041 "num_blocks": 65536, 00:14:15.041 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:15.041 "assigned_rate_limits": { 00:14:15.041 "rw_ios_per_sec": 0, 00:14:15.041 "rw_mbytes_per_sec": 0, 00:14:15.041 "r_mbytes_per_sec": 0, 00:14:15.041 "w_mbytes_per_sec": 0 00:14:15.041 }, 00:14:15.041 "claimed": false, 00:14:15.041 "zoned": false, 00:14:15.041 "supported_io_types": { 00:14:15.041 "read": true, 00:14:15.041 "write": true, 00:14:15.041 "unmap": true, 00:14:15.041 "flush": true, 00:14:15.041 "reset": true, 00:14:15.041 "nvme_admin": false, 00:14:15.041 "nvme_io": false, 00:14:15.041 "nvme_io_md": false, 00:14:15.041 "write_zeroes": true, 00:14:15.041 "zcopy": true, 00:14:15.041 "get_zone_info": false, 00:14:15.041 "zone_management": false, 00:14:15.041 "zone_append": false, 00:14:15.041 "compare": false, 00:14:15.041 "compare_and_write": false, 00:14:15.041 "abort": true, 00:14:15.041 "seek_hole": false, 00:14:15.041 "seek_data": false, 00:14:15.041 "copy": true, 00:14:15.041 "nvme_iov_md": false 00:14:15.041 }, 00:14:15.041 "memory_domains": [ 00:14:15.041 { 00:14:15.041 "dma_device_id": "system", 00:14:15.041 "dma_device_type": 1 00:14:15.041 }, 00:14:15.041 { 00:14:15.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:15.041 "dma_device_type": 2 00:14:15.041 } 00:14:15.041 ], 00:14:15.041 "driver_specific": {} 00:14:15.041 } 00:14:15.041 ] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.041 [2024-09-30 14:13:19.536793] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:15.041 [2024-09-30 14:13:19.536865] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:15.041 [2024-09-30 14:13:19.536883] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:15.041 [2024-09-30 14:13:19.538637] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:15.041 [2024-09-30 14:13:19.538683] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.041 "name": "Existed_Raid", 00:14:15.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.041 "strip_size_kb": 64, 00:14:15.041 "state": "configuring", 00:14:15.041 "raid_level": "raid5f", 00:14:15.041 "superblock": false, 00:14:15.041 "num_base_bdevs": 4, 00:14:15.041 "num_base_bdevs_discovered": 3, 00:14:15.041 "num_base_bdevs_operational": 4, 00:14:15.041 "base_bdevs_list": [ 00:14:15.041 { 00:14:15.041 "name": "BaseBdev1", 00:14:15.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.041 "is_configured": false, 00:14:15.041 "data_offset": 0, 00:14:15.041 "data_size": 0 00:14:15.041 }, 00:14:15.041 { 00:14:15.041 "name": "BaseBdev2", 00:14:15.041 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:15.041 "is_configured": true, 00:14:15.041 "data_offset": 0, 00:14:15.041 "data_size": 65536 00:14:15.041 }, 00:14:15.041 { 00:14:15.041 "name": "BaseBdev3", 00:14:15.041 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:15.041 "is_configured": true, 00:14:15.041 "data_offset": 0, 00:14:15.041 "data_size": 65536 00:14:15.041 }, 00:14:15.041 { 00:14:15.041 "name": "BaseBdev4", 00:14:15.041 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:15.041 "is_configured": true, 00:14:15.041 "data_offset": 0, 00:14:15.041 "data_size": 65536 00:14:15.041 } 00:14:15.041 ] 00:14:15.041 }' 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.041 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.606 [2024-09-30 14:13:19.964037] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.606 14:13:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.606 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.606 "name": "Existed_Raid", 00:14:15.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.606 "strip_size_kb": 64, 00:14:15.606 "state": "configuring", 00:14:15.606 "raid_level": "raid5f", 00:14:15.606 "superblock": false, 00:14:15.606 "num_base_bdevs": 4, 00:14:15.606 "num_base_bdevs_discovered": 2, 00:14:15.606 "num_base_bdevs_operational": 4, 00:14:15.606 "base_bdevs_list": [ 00:14:15.606 { 00:14:15.606 "name": "BaseBdev1", 00:14:15.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.606 "is_configured": false, 00:14:15.606 "data_offset": 0, 00:14:15.606 "data_size": 0 00:14:15.606 }, 00:14:15.606 { 00:14:15.606 "name": null, 00:14:15.606 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:15.606 "is_configured": false, 00:14:15.606 "data_offset": 0, 00:14:15.607 "data_size": 65536 00:14:15.607 }, 00:14:15.607 { 00:14:15.607 "name": "BaseBdev3", 00:14:15.607 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:15.607 "is_configured": true, 00:14:15.607 "data_offset": 0, 00:14:15.607 "data_size": 65536 00:14:15.607 }, 00:14:15.607 { 00:14:15.607 "name": "BaseBdev4", 00:14:15.607 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:15.607 "is_configured": true, 00:14:15.607 "data_offset": 0, 00:14:15.607 "data_size": 65536 00:14:15.607 } 00:14:15.607 ] 00:14:15.607 }' 00:14:15.607 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.607 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 [2024-09-30 14:13:20.442247] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:15.866 BaseBdev1 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 [ 00:14:15.866 { 00:14:15.866 "name": "BaseBdev1", 00:14:15.866 "aliases": [ 00:14:15.866 "76e78aeb-f7c0-4c1a-a919-23c26bcdc444" 00:14:15.866 ], 00:14:15.866 "product_name": "Malloc disk", 00:14:15.866 "block_size": 512, 00:14:15.866 "num_blocks": 65536, 00:14:15.866 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:15.866 "assigned_rate_limits": { 00:14:15.866 "rw_ios_per_sec": 0, 00:14:15.866 "rw_mbytes_per_sec": 0, 00:14:15.866 "r_mbytes_per_sec": 0, 00:14:15.866 "w_mbytes_per_sec": 0 00:14:15.866 }, 00:14:15.866 "claimed": true, 00:14:15.866 "claim_type": "exclusive_write", 00:14:15.866 "zoned": false, 00:14:15.866 "supported_io_types": { 00:14:15.866 "read": true, 00:14:15.866 "write": true, 00:14:15.866 "unmap": true, 00:14:15.866 "flush": true, 00:14:15.866 "reset": true, 00:14:15.866 "nvme_admin": false, 00:14:15.866 "nvme_io": false, 00:14:15.866 "nvme_io_md": false, 00:14:15.866 "write_zeroes": true, 00:14:15.866 "zcopy": true, 00:14:15.866 "get_zone_info": false, 00:14:15.866 "zone_management": false, 00:14:15.866 "zone_append": false, 00:14:15.866 "compare": false, 00:14:15.866 "compare_and_write": false, 00:14:15.866 "abort": true, 00:14:15.866 "seek_hole": false, 00:14:15.866 "seek_data": false, 00:14:15.866 "copy": true, 00:14:15.866 "nvme_iov_md": false 00:14:15.866 }, 00:14:15.866 "memory_domains": [ 00:14:15.866 { 00:14:15.866 "dma_device_id": "system", 00:14:15.866 "dma_device_type": 1 00:14:15.866 }, 00:14:15.866 { 00:14:15.866 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:15.866 "dma_device_type": 2 00:14:15.866 } 00:14:15.866 ], 00:14:15.866 "driver_specific": {} 00:14:15.866 } 00:14:15.866 ] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.866 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.125 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.125 "name": "Existed_Raid", 00:14:16.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:16.125 "strip_size_kb": 64, 00:14:16.125 "state": "configuring", 00:14:16.125 "raid_level": "raid5f", 00:14:16.125 "superblock": false, 00:14:16.125 "num_base_bdevs": 4, 00:14:16.125 "num_base_bdevs_discovered": 3, 00:14:16.125 "num_base_bdevs_operational": 4, 00:14:16.125 "base_bdevs_list": [ 00:14:16.125 { 00:14:16.125 "name": "BaseBdev1", 00:14:16.125 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:16.125 "is_configured": true, 00:14:16.125 "data_offset": 0, 00:14:16.125 "data_size": 65536 00:14:16.125 }, 00:14:16.125 { 00:14:16.125 "name": null, 00:14:16.125 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:16.125 "is_configured": false, 00:14:16.125 "data_offset": 0, 00:14:16.125 "data_size": 65536 00:14:16.125 }, 00:14:16.125 { 00:14:16.125 "name": "BaseBdev3", 00:14:16.125 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:16.125 "is_configured": true, 00:14:16.125 "data_offset": 0, 00:14:16.125 "data_size": 65536 00:14:16.125 }, 00:14:16.125 { 00:14:16.125 "name": "BaseBdev4", 00:14:16.125 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:16.125 "is_configured": true, 00:14:16.125 "data_offset": 0, 00:14:16.125 "data_size": 65536 00:14:16.125 } 00:14:16.125 ] 00:14:16.125 }' 00:14:16.125 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.125 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.384 [2024-09-30 14:13:20.953408] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.384 14:13:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.384 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.384 "name": "Existed_Raid", 00:14:16.384 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:16.384 "strip_size_kb": 64, 00:14:16.384 "state": "configuring", 00:14:16.384 "raid_level": "raid5f", 00:14:16.384 "superblock": false, 00:14:16.384 "num_base_bdevs": 4, 00:14:16.384 "num_base_bdevs_discovered": 2, 00:14:16.384 "num_base_bdevs_operational": 4, 00:14:16.384 "base_bdevs_list": [ 00:14:16.384 { 00:14:16.384 "name": "BaseBdev1", 00:14:16.384 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:16.384 "is_configured": true, 00:14:16.384 "data_offset": 0, 00:14:16.384 "data_size": 65536 00:14:16.384 }, 00:14:16.384 { 00:14:16.384 "name": null, 00:14:16.384 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:16.384 "is_configured": false, 00:14:16.384 "data_offset": 0, 00:14:16.384 "data_size": 65536 00:14:16.384 }, 00:14:16.384 { 00:14:16.384 "name": null, 00:14:16.384 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:16.384 "is_configured": false, 00:14:16.384 "data_offset": 0, 00:14:16.384 "data_size": 65536 00:14:16.384 }, 00:14:16.384 { 00:14:16.384 "name": "BaseBdev4", 00:14:16.384 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:16.384 "is_configured": true, 00:14:16.384 "data_offset": 0, 00:14:16.384 "data_size": 65536 00:14:16.384 } 00:14:16.384 ] 00:14:16.384 }' 00:14:16.384 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.384 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.951 [2024-09-30 14:13:21.468560] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.951 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.951 "name": "Existed_Raid", 00:14:16.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:16.951 "strip_size_kb": 64, 00:14:16.951 "state": "configuring", 00:14:16.951 "raid_level": "raid5f", 00:14:16.951 "superblock": false, 00:14:16.951 "num_base_bdevs": 4, 00:14:16.951 "num_base_bdevs_discovered": 3, 00:14:16.951 "num_base_bdevs_operational": 4, 00:14:16.951 "base_bdevs_list": [ 00:14:16.951 { 00:14:16.951 "name": "BaseBdev1", 00:14:16.951 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:16.951 "is_configured": true, 00:14:16.951 "data_offset": 0, 00:14:16.951 "data_size": 65536 00:14:16.951 }, 00:14:16.951 { 00:14:16.951 "name": null, 00:14:16.952 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:16.952 "is_configured": false, 00:14:16.952 "data_offset": 0, 00:14:16.952 "data_size": 65536 00:14:16.952 }, 00:14:16.952 { 00:14:16.952 "name": "BaseBdev3", 00:14:16.952 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:16.952 "is_configured": true, 00:14:16.952 "data_offset": 0, 00:14:16.952 "data_size": 65536 00:14:16.952 }, 00:14:16.952 { 00:14:16.952 "name": "BaseBdev4", 00:14:16.952 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:16.952 "is_configured": true, 00:14:16.952 "data_offset": 0, 00:14:16.952 "data_size": 65536 00:14:16.952 } 00:14:16.952 ] 00:14:16.952 }' 00:14:16.952 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.952 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.520 [2024-09-30 14:13:21.927769] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:17.520 "name": "Existed_Raid", 00:14:17.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:17.520 "strip_size_kb": 64, 00:14:17.520 "state": "configuring", 00:14:17.520 "raid_level": "raid5f", 00:14:17.520 "superblock": false, 00:14:17.520 "num_base_bdevs": 4, 00:14:17.520 "num_base_bdevs_discovered": 2, 00:14:17.520 "num_base_bdevs_operational": 4, 00:14:17.520 "base_bdevs_list": [ 00:14:17.520 { 00:14:17.520 "name": null, 00:14:17.520 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:17.520 "is_configured": false, 00:14:17.520 "data_offset": 0, 00:14:17.520 "data_size": 65536 00:14:17.520 }, 00:14:17.520 { 00:14:17.520 "name": null, 00:14:17.520 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:17.520 "is_configured": false, 00:14:17.520 "data_offset": 0, 00:14:17.520 "data_size": 65536 00:14:17.520 }, 00:14:17.520 { 00:14:17.520 "name": "BaseBdev3", 00:14:17.520 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:17.520 "is_configured": true, 00:14:17.520 "data_offset": 0, 00:14:17.520 "data_size": 65536 00:14:17.520 }, 00:14:17.520 { 00:14:17.520 "name": "BaseBdev4", 00:14:17.520 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:17.520 "is_configured": true, 00:14:17.520 "data_offset": 0, 00:14:17.520 "data_size": 65536 00:14:17.520 } 00:14:17.520 ] 00:14:17.520 }' 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:17.520 14:13:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.781 [2024-09-30 14:13:22.401461] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.781 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.040 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.040 "name": "Existed_Raid", 00:14:18.040 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.040 "strip_size_kb": 64, 00:14:18.040 "state": "configuring", 00:14:18.040 "raid_level": "raid5f", 00:14:18.040 "superblock": false, 00:14:18.040 "num_base_bdevs": 4, 00:14:18.040 "num_base_bdevs_discovered": 3, 00:14:18.040 "num_base_bdevs_operational": 4, 00:14:18.040 "base_bdevs_list": [ 00:14:18.040 { 00:14:18.040 "name": null, 00:14:18.040 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:18.040 "is_configured": false, 00:14:18.040 "data_offset": 0, 00:14:18.040 "data_size": 65536 00:14:18.040 }, 00:14:18.040 { 00:14:18.040 "name": "BaseBdev2", 00:14:18.040 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:18.040 "is_configured": true, 00:14:18.040 "data_offset": 0, 00:14:18.040 "data_size": 65536 00:14:18.040 }, 00:14:18.040 { 00:14:18.040 "name": "BaseBdev3", 00:14:18.040 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:18.040 "is_configured": true, 00:14:18.040 "data_offset": 0, 00:14:18.040 "data_size": 65536 00:14:18.040 }, 00:14:18.040 { 00:14:18.040 "name": "BaseBdev4", 00:14:18.040 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:18.040 "is_configured": true, 00:14:18.040 "data_offset": 0, 00:14:18.040 "data_size": 65536 00:14:18.040 } 00:14:18.040 ] 00:14:18.040 }' 00:14:18.040 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.040 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 76e78aeb-f7c0-4c1a-a919-23c26bcdc444 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 [2024-09-30 14:13:22.923460] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:18.299 [2024-09-30 14:13:22.923574] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:14:18.299 [2024-09-30 14:13:22.923589] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:18.299 [2024-09-30 14:13:22.923838] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:14:18.299 [2024-09-30 14:13:22.924266] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:14:18.299 [2024-09-30 14:13:22.924277] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:14:18.299 [2024-09-30 14:13:22.924441] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:18.299 NewBaseBdev 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.299 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.299 [ 00:14:18.299 { 00:14:18.299 "name": "NewBaseBdev", 00:14:18.299 "aliases": [ 00:14:18.299 "76e78aeb-f7c0-4c1a-a919-23c26bcdc444" 00:14:18.299 ], 00:14:18.299 "product_name": "Malloc disk", 00:14:18.299 "block_size": 512, 00:14:18.299 "num_blocks": 65536, 00:14:18.299 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:18.299 "assigned_rate_limits": { 00:14:18.299 "rw_ios_per_sec": 0, 00:14:18.299 "rw_mbytes_per_sec": 0, 00:14:18.299 "r_mbytes_per_sec": 0, 00:14:18.299 "w_mbytes_per_sec": 0 00:14:18.299 }, 00:14:18.299 "claimed": true, 00:14:18.299 "claim_type": "exclusive_write", 00:14:18.299 "zoned": false, 00:14:18.299 "supported_io_types": { 00:14:18.299 "read": true, 00:14:18.299 "write": true, 00:14:18.299 "unmap": true, 00:14:18.558 "flush": true, 00:14:18.558 "reset": true, 00:14:18.558 "nvme_admin": false, 00:14:18.558 "nvme_io": false, 00:14:18.558 "nvme_io_md": false, 00:14:18.558 "write_zeroes": true, 00:14:18.558 "zcopy": true, 00:14:18.558 "get_zone_info": false, 00:14:18.558 "zone_management": false, 00:14:18.558 "zone_append": false, 00:14:18.558 "compare": false, 00:14:18.558 "compare_and_write": false, 00:14:18.558 "abort": true, 00:14:18.558 "seek_hole": false, 00:14:18.558 "seek_data": false, 00:14:18.558 "copy": true, 00:14:18.558 "nvme_iov_md": false 00:14:18.558 }, 00:14:18.558 "memory_domains": [ 00:14:18.558 { 00:14:18.558 "dma_device_id": "system", 00:14:18.558 "dma_device_type": 1 00:14:18.558 }, 00:14:18.558 { 00:14:18.558 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:18.558 "dma_device_type": 2 00:14:18.558 } 00:14:18.558 ], 00:14:18.558 "driver_specific": {} 00:14:18.558 } 00:14:18.558 ] 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.558 14:13:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.558 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.558 "name": "Existed_Raid", 00:14:18.558 "uuid": "3f932b60-08c4-437f-b0f9-37bba2b55228", 00:14:18.558 "strip_size_kb": 64, 00:14:18.558 "state": "online", 00:14:18.558 "raid_level": "raid5f", 00:14:18.558 "superblock": false, 00:14:18.558 "num_base_bdevs": 4, 00:14:18.558 "num_base_bdevs_discovered": 4, 00:14:18.558 "num_base_bdevs_operational": 4, 00:14:18.558 "base_bdevs_list": [ 00:14:18.558 { 00:14:18.558 "name": "NewBaseBdev", 00:14:18.558 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:18.558 "is_configured": true, 00:14:18.558 "data_offset": 0, 00:14:18.558 "data_size": 65536 00:14:18.558 }, 00:14:18.558 { 00:14:18.558 "name": "BaseBdev2", 00:14:18.558 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:18.558 "is_configured": true, 00:14:18.558 "data_offset": 0, 00:14:18.558 "data_size": 65536 00:14:18.558 }, 00:14:18.558 { 00:14:18.558 "name": "BaseBdev3", 00:14:18.558 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:18.558 "is_configured": true, 00:14:18.558 "data_offset": 0, 00:14:18.558 "data_size": 65536 00:14:18.558 }, 00:14:18.558 { 00:14:18.558 "name": "BaseBdev4", 00:14:18.558 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:18.558 "is_configured": true, 00:14:18.558 "data_offset": 0, 00:14:18.558 "data_size": 65536 00:14:18.558 } 00:14:18.558 ] 00:14:18.558 }' 00:14:18.558 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.558 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.817 [2024-09-30 14:13:23.406835] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:18.817 "name": "Existed_Raid", 00:14:18.817 "aliases": [ 00:14:18.817 "3f932b60-08c4-437f-b0f9-37bba2b55228" 00:14:18.817 ], 00:14:18.817 "product_name": "Raid Volume", 00:14:18.817 "block_size": 512, 00:14:18.817 "num_blocks": 196608, 00:14:18.817 "uuid": "3f932b60-08c4-437f-b0f9-37bba2b55228", 00:14:18.817 "assigned_rate_limits": { 00:14:18.817 "rw_ios_per_sec": 0, 00:14:18.817 "rw_mbytes_per_sec": 0, 00:14:18.817 "r_mbytes_per_sec": 0, 00:14:18.817 "w_mbytes_per_sec": 0 00:14:18.817 }, 00:14:18.817 "claimed": false, 00:14:18.817 "zoned": false, 00:14:18.817 "supported_io_types": { 00:14:18.817 "read": true, 00:14:18.817 "write": true, 00:14:18.817 "unmap": false, 00:14:18.817 "flush": false, 00:14:18.817 "reset": true, 00:14:18.817 "nvme_admin": false, 00:14:18.817 "nvme_io": false, 00:14:18.817 "nvme_io_md": false, 00:14:18.817 "write_zeroes": true, 00:14:18.817 "zcopy": false, 00:14:18.817 "get_zone_info": false, 00:14:18.817 "zone_management": false, 00:14:18.817 "zone_append": false, 00:14:18.817 "compare": false, 00:14:18.817 "compare_and_write": false, 00:14:18.817 "abort": false, 00:14:18.817 "seek_hole": false, 00:14:18.817 "seek_data": false, 00:14:18.817 "copy": false, 00:14:18.817 "nvme_iov_md": false 00:14:18.817 }, 00:14:18.817 "driver_specific": { 00:14:18.817 "raid": { 00:14:18.817 "uuid": "3f932b60-08c4-437f-b0f9-37bba2b55228", 00:14:18.817 "strip_size_kb": 64, 00:14:18.817 "state": "online", 00:14:18.817 "raid_level": "raid5f", 00:14:18.817 "superblock": false, 00:14:18.817 "num_base_bdevs": 4, 00:14:18.817 "num_base_bdevs_discovered": 4, 00:14:18.817 "num_base_bdevs_operational": 4, 00:14:18.817 "base_bdevs_list": [ 00:14:18.817 { 00:14:18.817 "name": "NewBaseBdev", 00:14:18.817 "uuid": "76e78aeb-f7c0-4c1a-a919-23c26bcdc444", 00:14:18.817 "is_configured": true, 00:14:18.817 "data_offset": 0, 00:14:18.817 "data_size": 65536 00:14:18.817 }, 00:14:18.817 { 00:14:18.817 "name": "BaseBdev2", 00:14:18.817 "uuid": "41aa94a9-d1fe-4e0f-bbe9-55f75f8df3bf", 00:14:18.817 "is_configured": true, 00:14:18.817 "data_offset": 0, 00:14:18.817 "data_size": 65536 00:14:18.817 }, 00:14:18.817 { 00:14:18.817 "name": "BaseBdev3", 00:14:18.817 "uuid": "17db7850-ce13-4d99-9734-673e39eb206f", 00:14:18.817 "is_configured": true, 00:14:18.817 "data_offset": 0, 00:14:18.817 "data_size": 65536 00:14:18.817 }, 00:14:18.817 { 00:14:18.817 "name": "BaseBdev4", 00:14:18.817 "uuid": "8b3352b1-8d94-4cab-933d-b5602bc93b96", 00:14:18.817 "is_configured": true, 00:14:18.817 "data_offset": 0, 00:14:18.817 "data_size": 65536 00:14:18.817 } 00:14:18.817 ] 00:14:18.817 } 00:14:18.817 } 00:14:18.817 }' 00:14:18.817 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:19.076 BaseBdev2 00:14:19.076 BaseBdev3 00:14:19.076 BaseBdev4' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.076 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.335 [2024-09-30 14:13:23.742093] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:19.335 [2024-09-30 14:13:23.742119] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:19.335 [2024-09-30 14:13:23.742181] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:19.335 [2024-09-30 14:13:23.742411] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:19.335 [2024-09-30 14:13:23.742425] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.335 14:13:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 94264 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 94264 ']' 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 94264 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94264 00:14:19.336 killing process with pid 94264 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94264' 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 94264 00:14:19.336 [2024-09-30 14:13:23.786121] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:19.336 14:13:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 94264 00:14:19.336 [2024-09-30 14:13:23.825557] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:19.595 00:14:19.595 real 0m9.504s 00:14:19.595 user 0m16.245s 00:14:19.595 sys 0m2.101s 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:19.595 ************************************ 00:14:19.595 END TEST raid5f_state_function_test 00:14:19.595 ************************************ 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.595 14:13:24 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:19.595 14:13:24 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:19.595 14:13:24 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:19.595 14:13:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:19.595 ************************************ 00:14:19.595 START TEST raid5f_state_function_test_sb 00:14:19.595 ************************************ 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 true 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=94919 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 94919' 00:14:19.595 Process raid pid: 94919 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 94919 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 94919 ']' 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:19.595 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:19.595 14:13:24 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.855 [2024-09-30 14:13:24.262046] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:14:19.855 [2024-09-30 14:13:24.262382] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:19.855 [2024-09-30 14:13:24.400123] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:14:19.855 [2024-09-30 14:13:24.429415] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:19.855 [2024-09-30 14:13:24.476076] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:20.114 [2024-09-30 14:13:24.518600] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:20.114 [2024-09-30 14:13:24.518710] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.682 [2024-09-30 14:13:25.072011] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:20.682 [2024-09-30 14:13:25.072063] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:20.682 [2024-09-30 14:13:25.072075] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:20.682 [2024-09-30 14:13:25.072082] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:20.682 [2024-09-30 14:13:25.072093] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:20.682 [2024-09-30 14:13:25.072100] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:20.682 [2024-09-30 14:13:25.072107] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:20.682 [2024-09-30 14:13:25.072114] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.682 "name": "Existed_Raid", 00:14:20.682 "uuid": "73b730dd-3e8c-49bb-a436-bbbbda3219c6", 00:14:20.682 "strip_size_kb": 64, 00:14:20.682 "state": "configuring", 00:14:20.682 "raid_level": "raid5f", 00:14:20.682 "superblock": true, 00:14:20.682 "num_base_bdevs": 4, 00:14:20.682 "num_base_bdevs_discovered": 0, 00:14:20.682 "num_base_bdevs_operational": 4, 00:14:20.682 "base_bdevs_list": [ 00:14:20.682 { 00:14:20.682 "name": "BaseBdev1", 00:14:20.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.682 "is_configured": false, 00:14:20.682 "data_offset": 0, 00:14:20.682 "data_size": 0 00:14:20.682 }, 00:14:20.682 { 00:14:20.682 "name": "BaseBdev2", 00:14:20.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.682 "is_configured": false, 00:14:20.682 "data_offset": 0, 00:14:20.682 "data_size": 0 00:14:20.682 }, 00:14:20.682 { 00:14:20.682 "name": "BaseBdev3", 00:14:20.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.682 "is_configured": false, 00:14:20.682 "data_offset": 0, 00:14:20.682 "data_size": 0 00:14:20.682 }, 00:14:20.682 { 00:14:20.682 "name": "BaseBdev4", 00:14:20.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.682 "is_configured": false, 00:14:20.682 "data_offset": 0, 00:14:20.682 "data_size": 0 00:14:20.682 } 00:14:20.682 ] 00:14:20.682 }' 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.682 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.941 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 [2024-09-30 14:13:25.479377] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:20.942 [2024-09-30 14:13:25.479489] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 [2024-09-30 14:13:25.491380] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:20.942 [2024-09-30 14:13:25.491455] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:20.942 [2024-09-30 14:13:25.491469] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:20.942 [2024-09-30 14:13:25.491476] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:20.942 [2024-09-30 14:13:25.491498] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:20.942 [2024-09-30 14:13:25.491505] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:20.942 [2024-09-30 14:13:25.491512] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:20.942 [2024-09-30 14:13:25.491519] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 [2024-09-30 14:13:25.512138] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:20.942 BaseBdev1 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 [ 00:14:20.942 { 00:14:20.942 "name": "BaseBdev1", 00:14:20.942 "aliases": [ 00:14:20.942 "62561fbe-6553-422c-9ea8-0a72f02c6491" 00:14:20.942 ], 00:14:20.942 "product_name": "Malloc disk", 00:14:20.942 "block_size": 512, 00:14:20.942 "num_blocks": 65536, 00:14:20.942 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:20.942 "assigned_rate_limits": { 00:14:20.942 "rw_ios_per_sec": 0, 00:14:20.942 "rw_mbytes_per_sec": 0, 00:14:20.942 "r_mbytes_per_sec": 0, 00:14:20.942 "w_mbytes_per_sec": 0 00:14:20.942 }, 00:14:20.942 "claimed": true, 00:14:20.942 "claim_type": "exclusive_write", 00:14:20.942 "zoned": false, 00:14:20.942 "supported_io_types": { 00:14:20.942 "read": true, 00:14:20.942 "write": true, 00:14:20.942 "unmap": true, 00:14:20.942 "flush": true, 00:14:20.942 "reset": true, 00:14:20.942 "nvme_admin": false, 00:14:20.942 "nvme_io": false, 00:14:20.942 "nvme_io_md": false, 00:14:20.942 "write_zeroes": true, 00:14:20.942 "zcopy": true, 00:14:20.942 "get_zone_info": false, 00:14:20.942 "zone_management": false, 00:14:20.942 "zone_append": false, 00:14:20.942 "compare": false, 00:14:20.942 "compare_and_write": false, 00:14:20.942 "abort": true, 00:14:20.942 "seek_hole": false, 00:14:20.942 "seek_data": false, 00:14:20.942 "copy": true, 00:14:20.942 "nvme_iov_md": false 00:14:20.942 }, 00:14:20.942 "memory_domains": [ 00:14:20.942 { 00:14:20.942 "dma_device_id": "system", 00:14:20.942 "dma_device_type": 1 00:14:20.942 }, 00:14:20.942 { 00:14:20.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:20.942 "dma_device_type": 2 00:14:20.942 } 00:14:20.942 ], 00:14:20.942 "driver_specific": {} 00:14:20.942 } 00:14:20.942 ] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.942 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.201 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.201 "name": "Existed_Raid", 00:14:21.201 "uuid": "40ff0fc8-baba-4f8f-92af-4210071f0564", 00:14:21.201 "strip_size_kb": 64, 00:14:21.201 "state": "configuring", 00:14:21.201 "raid_level": "raid5f", 00:14:21.201 "superblock": true, 00:14:21.201 "num_base_bdevs": 4, 00:14:21.201 "num_base_bdevs_discovered": 1, 00:14:21.201 "num_base_bdevs_operational": 4, 00:14:21.201 "base_bdevs_list": [ 00:14:21.201 { 00:14:21.201 "name": "BaseBdev1", 00:14:21.201 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:21.201 "is_configured": true, 00:14:21.201 "data_offset": 2048, 00:14:21.201 "data_size": 63488 00:14:21.201 }, 00:14:21.201 { 00:14:21.201 "name": "BaseBdev2", 00:14:21.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.201 "is_configured": false, 00:14:21.201 "data_offset": 0, 00:14:21.201 "data_size": 0 00:14:21.201 }, 00:14:21.201 { 00:14:21.201 "name": "BaseBdev3", 00:14:21.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.201 "is_configured": false, 00:14:21.201 "data_offset": 0, 00:14:21.201 "data_size": 0 00:14:21.201 }, 00:14:21.201 { 00:14:21.201 "name": "BaseBdev4", 00:14:21.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.201 "is_configured": false, 00:14:21.201 "data_offset": 0, 00:14:21.201 "data_size": 0 00:14:21.201 } 00:14:21.201 ] 00:14:21.201 }' 00:14:21.201 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.201 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.462 [2024-09-30 14:13:25.979388] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:21.462 [2024-09-30 14:13:25.979509] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.462 [2024-09-30 14:13:25.991408] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:21.462 [2024-09-30 14:13:25.993203] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:21.462 [2024-09-30 14:13:25.993271] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:21.462 [2024-09-30 14:13:25.993316] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:21.462 [2024-09-30 14:13:25.993335] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:21.462 [2024-09-30 14:13:25.993354] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:21.462 [2024-09-30 14:13:25.993372] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.462 14:13:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.462 "name": "Existed_Raid", 00:14:21.462 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:21.462 "strip_size_kb": 64, 00:14:21.462 "state": "configuring", 00:14:21.462 "raid_level": "raid5f", 00:14:21.462 "superblock": true, 00:14:21.462 "num_base_bdevs": 4, 00:14:21.462 "num_base_bdevs_discovered": 1, 00:14:21.462 "num_base_bdevs_operational": 4, 00:14:21.462 "base_bdevs_list": [ 00:14:21.462 { 00:14:21.462 "name": "BaseBdev1", 00:14:21.462 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:21.462 "is_configured": true, 00:14:21.462 "data_offset": 2048, 00:14:21.462 "data_size": 63488 00:14:21.462 }, 00:14:21.462 { 00:14:21.462 "name": "BaseBdev2", 00:14:21.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.462 "is_configured": false, 00:14:21.462 "data_offset": 0, 00:14:21.462 "data_size": 0 00:14:21.462 }, 00:14:21.462 { 00:14:21.462 "name": "BaseBdev3", 00:14:21.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.462 "is_configured": false, 00:14:21.462 "data_offset": 0, 00:14:21.462 "data_size": 0 00:14:21.462 }, 00:14:21.462 { 00:14:21.462 "name": "BaseBdev4", 00:14:21.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.462 "is_configured": false, 00:14:21.462 "data_offset": 0, 00:14:21.462 "data_size": 0 00:14:21.462 } 00:14:21.462 ] 00:14:21.462 }' 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.462 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.065 [2024-09-30 14:13:26.429114] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:22.065 BaseBdev2 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.065 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.065 [ 00:14:22.065 { 00:14:22.065 "name": "BaseBdev2", 00:14:22.065 "aliases": [ 00:14:22.065 "7026564e-52c9-4892-be44-c76abd18e7f4" 00:14:22.065 ], 00:14:22.065 "product_name": "Malloc disk", 00:14:22.065 "block_size": 512, 00:14:22.065 "num_blocks": 65536, 00:14:22.065 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:22.065 "assigned_rate_limits": { 00:14:22.065 "rw_ios_per_sec": 0, 00:14:22.065 "rw_mbytes_per_sec": 0, 00:14:22.065 "r_mbytes_per_sec": 0, 00:14:22.065 "w_mbytes_per_sec": 0 00:14:22.065 }, 00:14:22.065 "claimed": true, 00:14:22.065 "claim_type": "exclusive_write", 00:14:22.065 "zoned": false, 00:14:22.065 "supported_io_types": { 00:14:22.065 "read": true, 00:14:22.065 "write": true, 00:14:22.065 "unmap": true, 00:14:22.065 "flush": true, 00:14:22.065 "reset": true, 00:14:22.065 "nvme_admin": false, 00:14:22.065 "nvme_io": false, 00:14:22.065 "nvme_io_md": false, 00:14:22.065 "write_zeroes": true, 00:14:22.065 "zcopy": true, 00:14:22.065 "get_zone_info": false, 00:14:22.065 "zone_management": false, 00:14:22.065 "zone_append": false, 00:14:22.065 "compare": false, 00:14:22.065 "compare_and_write": false, 00:14:22.065 "abort": true, 00:14:22.065 "seek_hole": false, 00:14:22.065 "seek_data": false, 00:14:22.065 "copy": true, 00:14:22.065 "nvme_iov_md": false 00:14:22.065 }, 00:14:22.066 "memory_domains": [ 00:14:22.066 { 00:14:22.066 "dma_device_id": "system", 00:14:22.066 "dma_device_type": 1 00:14:22.066 }, 00:14:22.066 { 00:14:22.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.066 "dma_device_type": 2 00:14:22.066 } 00:14:22.066 ], 00:14:22.066 "driver_specific": {} 00:14:22.066 } 00:14:22.066 ] 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.066 "name": "Existed_Raid", 00:14:22.066 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:22.066 "strip_size_kb": 64, 00:14:22.066 "state": "configuring", 00:14:22.066 "raid_level": "raid5f", 00:14:22.066 "superblock": true, 00:14:22.066 "num_base_bdevs": 4, 00:14:22.066 "num_base_bdevs_discovered": 2, 00:14:22.066 "num_base_bdevs_operational": 4, 00:14:22.066 "base_bdevs_list": [ 00:14:22.066 { 00:14:22.066 "name": "BaseBdev1", 00:14:22.066 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:22.066 "is_configured": true, 00:14:22.066 "data_offset": 2048, 00:14:22.066 "data_size": 63488 00:14:22.066 }, 00:14:22.066 { 00:14:22.066 "name": "BaseBdev2", 00:14:22.066 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:22.066 "is_configured": true, 00:14:22.066 "data_offset": 2048, 00:14:22.066 "data_size": 63488 00:14:22.066 }, 00:14:22.066 { 00:14:22.066 "name": "BaseBdev3", 00:14:22.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.066 "is_configured": false, 00:14:22.066 "data_offset": 0, 00:14:22.066 "data_size": 0 00:14:22.066 }, 00:14:22.066 { 00:14:22.066 "name": "BaseBdev4", 00:14:22.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.066 "is_configured": false, 00:14:22.066 "data_offset": 0, 00:14:22.066 "data_size": 0 00:14:22.066 } 00:14:22.066 ] 00:14:22.066 }' 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.066 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.334 [2024-09-30 14:13:26.911199] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:22.334 BaseBdev3 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.334 [ 00:14:22.334 { 00:14:22.334 "name": "BaseBdev3", 00:14:22.334 "aliases": [ 00:14:22.334 "68e03f23-0f3d-473c-b58c-5f52e99335f4" 00:14:22.334 ], 00:14:22.334 "product_name": "Malloc disk", 00:14:22.334 "block_size": 512, 00:14:22.334 "num_blocks": 65536, 00:14:22.334 "uuid": "68e03f23-0f3d-473c-b58c-5f52e99335f4", 00:14:22.334 "assigned_rate_limits": { 00:14:22.334 "rw_ios_per_sec": 0, 00:14:22.334 "rw_mbytes_per_sec": 0, 00:14:22.334 "r_mbytes_per_sec": 0, 00:14:22.334 "w_mbytes_per_sec": 0 00:14:22.334 }, 00:14:22.334 "claimed": true, 00:14:22.334 "claim_type": "exclusive_write", 00:14:22.334 "zoned": false, 00:14:22.334 "supported_io_types": { 00:14:22.334 "read": true, 00:14:22.334 "write": true, 00:14:22.334 "unmap": true, 00:14:22.334 "flush": true, 00:14:22.334 "reset": true, 00:14:22.334 "nvme_admin": false, 00:14:22.334 "nvme_io": false, 00:14:22.334 "nvme_io_md": false, 00:14:22.334 "write_zeroes": true, 00:14:22.334 "zcopy": true, 00:14:22.334 "get_zone_info": false, 00:14:22.334 "zone_management": false, 00:14:22.334 "zone_append": false, 00:14:22.334 "compare": false, 00:14:22.334 "compare_and_write": false, 00:14:22.334 "abort": true, 00:14:22.334 "seek_hole": false, 00:14:22.334 "seek_data": false, 00:14:22.334 "copy": true, 00:14:22.334 "nvme_iov_md": false 00:14:22.334 }, 00:14:22.334 "memory_domains": [ 00:14:22.334 { 00:14:22.334 "dma_device_id": "system", 00:14:22.334 "dma_device_type": 1 00:14:22.334 }, 00:14:22.334 { 00:14:22.334 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.334 "dma_device_type": 2 00:14:22.334 } 00:14:22.334 ], 00:14:22.334 "driver_specific": {} 00:14:22.334 } 00:14:22.334 ] 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:22.334 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.335 14:13:26 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.600 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.600 "name": "Existed_Raid", 00:14:22.600 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:22.600 "strip_size_kb": 64, 00:14:22.600 "state": "configuring", 00:14:22.600 "raid_level": "raid5f", 00:14:22.600 "superblock": true, 00:14:22.600 "num_base_bdevs": 4, 00:14:22.600 "num_base_bdevs_discovered": 3, 00:14:22.600 "num_base_bdevs_operational": 4, 00:14:22.600 "base_bdevs_list": [ 00:14:22.600 { 00:14:22.600 "name": "BaseBdev1", 00:14:22.600 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:22.600 "is_configured": true, 00:14:22.600 "data_offset": 2048, 00:14:22.600 "data_size": 63488 00:14:22.600 }, 00:14:22.600 { 00:14:22.600 "name": "BaseBdev2", 00:14:22.600 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:22.600 "is_configured": true, 00:14:22.600 "data_offset": 2048, 00:14:22.600 "data_size": 63488 00:14:22.600 }, 00:14:22.600 { 00:14:22.600 "name": "BaseBdev3", 00:14:22.600 "uuid": "68e03f23-0f3d-473c-b58c-5f52e99335f4", 00:14:22.600 "is_configured": true, 00:14:22.600 "data_offset": 2048, 00:14:22.600 "data_size": 63488 00:14:22.600 }, 00:14:22.600 { 00:14:22.600 "name": "BaseBdev4", 00:14:22.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.600 "is_configured": false, 00:14:22.600 "data_offset": 0, 00:14:22.600 "data_size": 0 00:14:22.600 } 00:14:22.600 ] 00:14:22.600 }' 00:14:22.600 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.600 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.860 [2024-09-30 14:13:27.341507] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:22.860 [2024-09-30 14:13:27.341702] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:14:22.860 [2024-09-30 14:13:27.341718] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:22.860 [2024-09-30 14:13:27.341971] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:22.860 BaseBdev4 00:14:22.860 [2024-09-30 14:13:27.342410] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:14:22.860 [2024-09-30 14:13:27.342428] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:14:22.860 [2024-09-30 14:13:27.342579] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.860 [ 00:14:22.860 { 00:14:22.860 "name": "BaseBdev4", 00:14:22.860 "aliases": [ 00:14:22.860 "0c1bc6a2-851b-4645-a541-d369c4aff0d3" 00:14:22.860 ], 00:14:22.860 "product_name": "Malloc disk", 00:14:22.860 "block_size": 512, 00:14:22.860 "num_blocks": 65536, 00:14:22.860 "uuid": "0c1bc6a2-851b-4645-a541-d369c4aff0d3", 00:14:22.860 "assigned_rate_limits": { 00:14:22.860 "rw_ios_per_sec": 0, 00:14:22.860 "rw_mbytes_per_sec": 0, 00:14:22.860 "r_mbytes_per_sec": 0, 00:14:22.860 "w_mbytes_per_sec": 0 00:14:22.860 }, 00:14:22.860 "claimed": true, 00:14:22.860 "claim_type": "exclusive_write", 00:14:22.860 "zoned": false, 00:14:22.860 "supported_io_types": { 00:14:22.860 "read": true, 00:14:22.860 "write": true, 00:14:22.860 "unmap": true, 00:14:22.860 "flush": true, 00:14:22.860 "reset": true, 00:14:22.860 "nvme_admin": false, 00:14:22.860 "nvme_io": false, 00:14:22.860 "nvme_io_md": false, 00:14:22.860 "write_zeroes": true, 00:14:22.860 "zcopy": true, 00:14:22.860 "get_zone_info": false, 00:14:22.860 "zone_management": false, 00:14:22.860 "zone_append": false, 00:14:22.860 "compare": false, 00:14:22.860 "compare_and_write": false, 00:14:22.860 "abort": true, 00:14:22.860 "seek_hole": false, 00:14:22.860 "seek_data": false, 00:14:22.860 "copy": true, 00:14:22.860 "nvme_iov_md": false 00:14:22.860 }, 00:14:22.860 "memory_domains": [ 00:14:22.860 { 00:14:22.860 "dma_device_id": "system", 00:14:22.860 "dma_device_type": 1 00:14:22.860 }, 00:14:22.860 { 00:14:22.860 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.860 "dma_device_type": 2 00:14:22.860 } 00:14:22.860 ], 00:14:22.860 "driver_specific": {} 00:14:22.860 } 00:14:22.860 ] 00:14:22.860 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.861 "name": "Existed_Raid", 00:14:22.861 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:22.861 "strip_size_kb": 64, 00:14:22.861 "state": "online", 00:14:22.861 "raid_level": "raid5f", 00:14:22.861 "superblock": true, 00:14:22.861 "num_base_bdevs": 4, 00:14:22.861 "num_base_bdevs_discovered": 4, 00:14:22.861 "num_base_bdevs_operational": 4, 00:14:22.861 "base_bdevs_list": [ 00:14:22.861 { 00:14:22.861 "name": "BaseBdev1", 00:14:22.861 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:22.861 "is_configured": true, 00:14:22.861 "data_offset": 2048, 00:14:22.861 "data_size": 63488 00:14:22.861 }, 00:14:22.861 { 00:14:22.861 "name": "BaseBdev2", 00:14:22.861 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:22.861 "is_configured": true, 00:14:22.861 "data_offset": 2048, 00:14:22.861 "data_size": 63488 00:14:22.861 }, 00:14:22.861 { 00:14:22.861 "name": "BaseBdev3", 00:14:22.861 "uuid": "68e03f23-0f3d-473c-b58c-5f52e99335f4", 00:14:22.861 "is_configured": true, 00:14:22.861 "data_offset": 2048, 00:14:22.861 "data_size": 63488 00:14:22.861 }, 00:14:22.861 { 00:14:22.861 "name": "BaseBdev4", 00:14:22.861 "uuid": "0c1bc6a2-851b-4645-a541-d369c4aff0d3", 00:14:22.861 "is_configured": true, 00:14:22.861 "data_offset": 2048, 00:14:22.861 "data_size": 63488 00:14:22.861 } 00:14:22.861 ] 00:14:22.861 }' 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.861 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.431 [2024-09-30 14:13:27.824878] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:23.431 "name": "Existed_Raid", 00:14:23.431 "aliases": [ 00:14:23.431 "71cb63fd-9f02-4cf5-b70d-193ea63e39bc" 00:14:23.431 ], 00:14:23.431 "product_name": "Raid Volume", 00:14:23.431 "block_size": 512, 00:14:23.431 "num_blocks": 190464, 00:14:23.431 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:23.431 "assigned_rate_limits": { 00:14:23.431 "rw_ios_per_sec": 0, 00:14:23.431 "rw_mbytes_per_sec": 0, 00:14:23.431 "r_mbytes_per_sec": 0, 00:14:23.431 "w_mbytes_per_sec": 0 00:14:23.431 }, 00:14:23.431 "claimed": false, 00:14:23.431 "zoned": false, 00:14:23.431 "supported_io_types": { 00:14:23.431 "read": true, 00:14:23.431 "write": true, 00:14:23.431 "unmap": false, 00:14:23.431 "flush": false, 00:14:23.431 "reset": true, 00:14:23.431 "nvme_admin": false, 00:14:23.431 "nvme_io": false, 00:14:23.431 "nvme_io_md": false, 00:14:23.431 "write_zeroes": true, 00:14:23.431 "zcopy": false, 00:14:23.431 "get_zone_info": false, 00:14:23.431 "zone_management": false, 00:14:23.431 "zone_append": false, 00:14:23.431 "compare": false, 00:14:23.431 "compare_and_write": false, 00:14:23.431 "abort": false, 00:14:23.431 "seek_hole": false, 00:14:23.431 "seek_data": false, 00:14:23.431 "copy": false, 00:14:23.431 "nvme_iov_md": false 00:14:23.431 }, 00:14:23.431 "driver_specific": { 00:14:23.431 "raid": { 00:14:23.431 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:23.431 "strip_size_kb": 64, 00:14:23.431 "state": "online", 00:14:23.431 "raid_level": "raid5f", 00:14:23.431 "superblock": true, 00:14:23.431 "num_base_bdevs": 4, 00:14:23.431 "num_base_bdevs_discovered": 4, 00:14:23.431 "num_base_bdevs_operational": 4, 00:14:23.431 "base_bdevs_list": [ 00:14:23.431 { 00:14:23.431 "name": "BaseBdev1", 00:14:23.431 "uuid": "62561fbe-6553-422c-9ea8-0a72f02c6491", 00:14:23.431 "is_configured": true, 00:14:23.431 "data_offset": 2048, 00:14:23.431 "data_size": 63488 00:14:23.431 }, 00:14:23.431 { 00:14:23.431 "name": "BaseBdev2", 00:14:23.431 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:23.431 "is_configured": true, 00:14:23.431 "data_offset": 2048, 00:14:23.431 "data_size": 63488 00:14:23.431 }, 00:14:23.431 { 00:14:23.431 "name": "BaseBdev3", 00:14:23.431 "uuid": "68e03f23-0f3d-473c-b58c-5f52e99335f4", 00:14:23.431 "is_configured": true, 00:14:23.431 "data_offset": 2048, 00:14:23.431 "data_size": 63488 00:14:23.431 }, 00:14:23.431 { 00:14:23.431 "name": "BaseBdev4", 00:14:23.431 "uuid": "0c1bc6a2-851b-4645-a541-d369c4aff0d3", 00:14:23.431 "is_configured": true, 00:14:23.431 "data_offset": 2048, 00:14:23.431 "data_size": 63488 00:14:23.431 } 00:14:23.431 ] 00:14:23.431 } 00:14:23.431 } 00:14:23.431 }' 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:23.431 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:23.431 BaseBdev2 00:14:23.431 BaseBdev3 00:14:23.431 BaseBdev4' 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.432 14:13:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.432 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.692 [2024-09-30 14:13:28.176146] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.692 "name": "Existed_Raid", 00:14:23.692 "uuid": "71cb63fd-9f02-4cf5-b70d-193ea63e39bc", 00:14:23.692 "strip_size_kb": 64, 00:14:23.692 "state": "online", 00:14:23.692 "raid_level": "raid5f", 00:14:23.692 "superblock": true, 00:14:23.692 "num_base_bdevs": 4, 00:14:23.692 "num_base_bdevs_discovered": 3, 00:14:23.692 "num_base_bdevs_operational": 3, 00:14:23.692 "base_bdevs_list": [ 00:14:23.692 { 00:14:23.692 "name": null, 00:14:23.692 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.692 "is_configured": false, 00:14:23.692 "data_offset": 0, 00:14:23.692 "data_size": 63488 00:14:23.692 }, 00:14:23.692 { 00:14:23.692 "name": "BaseBdev2", 00:14:23.692 "uuid": "7026564e-52c9-4892-be44-c76abd18e7f4", 00:14:23.692 "is_configured": true, 00:14:23.692 "data_offset": 2048, 00:14:23.692 "data_size": 63488 00:14:23.692 }, 00:14:23.692 { 00:14:23.692 "name": "BaseBdev3", 00:14:23.692 "uuid": "68e03f23-0f3d-473c-b58c-5f52e99335f4", 00:14:23.692 "is_configured": true, 00:14:23.692 "data_offset": 2048, 00:14:23.692 "data_size": 63488 00:14:23.692 }, 00:14:23.692 { 00:14:23.692 "name": "BaseBdev4", 00:14:23.692 "uuid": "0c1bc6a2-851b-4645-a541-d369c4aff0d3", 00:14:23.692 "is_configured": true, 00:14:23.692 "data_offset": 2048, 00:14:23.692 "data_size": 63488 00:14:23.692 } 00:14:23.692 ] 00:14:23.692 }' 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.692 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.262 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 [2024-09-30 14:13:28.706677] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:24.263 [2024-09-30 14:13:28.706859] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:24.263 [2024-09-30 14:13:28.717760] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 [2024-09-30 14:13:28.765720] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 [2024-09-30 14:13:28.836308] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:24.263 [2024-09-30 14:13:28.836394] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.263 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.263 BaseBdev2 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.527 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.527 [ 00:14:24.527 { 00:14:24.527 "name": "BaseBdev2", 00:14:24.527 "aliases": [ 00:14:24.527 "639f25a8-a4b1-4263-8dbf-43592e64a60f" 00:14:24.527 ], 00:14:24.527 "product_name": "Malloc disk", 00:14:24.527 "block_size": 512, 00:14:24.527 "num_blocks": 65536, 00:14:24.527 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:24.527 "assigned_rate_limits": { 00:14:24.527 "rw_ios_per_sec": 0, 00:14:24.527 "rw_mbytes_per_sec": 0, 00:14:24.527 "r_mbytes_per_sec": 0, 00:14:24.527 "w_mbytes_per_sec": 0 00:14:24.527 }, 00:14:24.527 "claimed": false, 00:14:24.527 "zoned": false, 00:14:24.527 "supported_io_types": { 00:14:24.527 "read": true, 00:14:24.527 "write": true, 00:14:24.527 "unmap": true, 00:14:24.527 "flush": true, 00:14:24.527 "reset": true, 00:14:24.527 "nvme_admin": false, 00:14:24.527 "nvme_io": false, 00:14:24.527 "nvme_io_md": false, 00:14:24.527 "write_zeroes": true, 00:14:24.527 "zcopy": true, 00:14:24.527 "get_zone_info": false, 00:14:24.527 "zone_management": false, 00:14:24.527 "zone_append": false, 00:14:24.527 "compare": false, 00:14:24.527 "compare_and_write": false, 00:14:24.527 "abort": true, 00:14:24.527 "seek_hole": false, 00:14:24.527 "seek_data": false, 00:14:24.528 "copy": true, 00:14:24.528 "nvme_iov_md": false 00:14:24.528 }, 00:14:24.528 "memory_domains": [ 00:14:24.528 { 00:14:24.528 "dma_device_id": "system", 00:14:24.528 "dma_device_type": 1 00:14:24.528 }, 00:14:24.528 { 00:14:24.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.528 "dma_device_type": 2 00:14:24.528 } 00:14:24.528 ], 00:14:24.528 "driver_specific": {} 00:14:24.528 } 00:14:24.528 ] 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.528 BaseBdev3 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.528 14:13:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.528 [ 00:14:24.528 { 00:14:24.528 "name": "BaseBdev3", 00:14:24.528 "aliases": [ 00:14:24.528 "eb2a6e28-6d70-4515-895f-c6cefe6e441d" 00:14:24.528 ], 00:14:24.528 "product_name": "Malloc disk", 00:14:24.528 "block_size": 512, 00:14:24.528 "num_blocks": 65536, 00:14:24.528 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:24.528 "assigned_rate_limits": { 00:14:24.528 "rw_ios_per_sec": 0, 00:14:24.528 "rw_mbytes_per_sec": 0, 00:14:24.528 "r_mbytes_per_sec": 0, 00:14:24.528 "w_mbytes_per_sec": 0 00:14:24.528 }, 00:14:24.528 "claimed": false, 00:14:24.528 "zoned": false, 00:14:24.528 "supported_io_types": { 00:14:24.528 "read": true, 00:14:24.528 "write": true, 00:14:24.528 "unmap": true, 00:14:24.528 "flush": true, 00:14:24.528 "reset": true, 00:14:24.528 "nvme_admin": false, 00:14:24.528 "nvme_io": false, 00:14:24.528 "nvme_io_md": false, 00:14:24.528 "write_zeroes": true, 00:14:24.528 "zcopy": true, 00:14:24.528 "get_zone_info": false, 00:14:24.528 "zone_management": false, 00:14:24.528 "zone_append": false, 00:14:24.528 "compare": false, 00:14:24.528 "compare_and_write": false, 00:14:24.528 "abort": true, 00:14:24.528 "seek_hole": false, 00:14:24.528 "seek_data": false, 00:14:24.528 "copy": true, 00:14:24.528 "nvme_iov_md": false 00:14:24.528 }, 00:14:24.528 "memory_domains": [ 00:14:24.528 { 00:14:24.528 "dma_device_id": "system", 00:14:24.528 "dma_device_type": 1 00:14:24.528 }, 00:14:24.528 { 00:14:24.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.528 "dma_device_type": 2 00:14:24.528 } 00:14:24.528 ], 00:14:24.528 "driver_specific": {} 00:14:24.528 } 00:14:24.528 ] 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:24.528 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.529 BaseBdev4 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.529 [ 00:14:24.529 { 00:14:24.529 "name": "BaseBdev4", 00:14:24.529 "aliases": [ 00:14:24.529 "a276601c-fec4-4f2f-805a-7e218f02d601" 00:14:24.529 ], 00:14:24.529 "product_name": "Malloc disk", 00:14:24.529 "block_size": 512, 00:14:24.529 "num_blocks": 65536, 00:14:24.529 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:24.529 "assigned_rate_limits": { 00:14:24.529 "rw_ios_per_sec": 0, 00:14:24.529 "rw_mbytes_per_sec": 0, 00:14:24.529 "r_mbytes_per_sec": 0, 00:14:24.529 "w_mbytes_per_sec": 0 00:14:24.529 }, 00:14:24.529 "claimed": false, 00:14:24.529 "zoned": false, 00:14:24.529 "supported_io_types": { 00:14:24.529 "read": true, 00:14:24.529 "write": true, 00:14:24.529 "unmap": true, 00:14:24.529 "flush": true, 00:14:24.529 "reset": true, 00:14:24.529 "nvme_admin": false, 00:14:24.529 "nvme_io": false, 00:14:24.529 "nvme_io_md": false, 00:14:24.529 "write_zeroes": true, 00:14:24.529 "zcopy": true, 00:14:24.529 "get_zone_info": false, 00:14:24.529 "zone_management": false, 00:14:24.529 "zone_append": false, 00:14:24.529 "compare": false, 00:14:24.529 "compare_and_write": false, 00:14:24.529 "abort": true, 00:14:24.529 "seek_hole": false, 00:14:24.529 "seek_data": false, 00:14:24.529 "copy": true, 00:14:24.529 "nvme_iov_md": false 00:14:24.529 }, 00:14:24.529 "memory_domains": [ 00:14:24.529 { 00:14:24.529 "dma_device_id": "system", 00:14:24.529 "dma_device_type": 1 00:14:24.529 }, 00:14:24.529 { 00:14:24.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.529 "dma_device_type": 2 00:14:24.529 } 00:14:24.529 ], 00:14:24.529 "driver_specific": {} 00:14:24.529 } 00:14:24.529 ] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.529 [2024-09-30 14:13:29.066720] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:24.529 [2024-09-30 14:13:29.066835] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:24.529 [2024-09-30 14:13:29.066878] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:24.529 [2024-09-30 14:13:29.068664] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:24.529 [2024-09-30 14:13:29.068748] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.529 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.530 "name": "Existed_Raid", 00:14:24.530 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:24.530 "strip_size_kb": 64, 00:14:24.530 "state": "configuring", 00:14:24.530 "raid_level": "raid5f", 00:14:24.530 "superblock": true, 00:14:24.530 "num_base_bdevs": 4, 00:14:24.530 "num_base_bdevs_discovered": 3, 00:14:24.530 "num_base_bdevs_operational": 4, 00:14:24.530 "base_bdevs_list": [ 00:14:24.530 { 00:14:24.530 "name": "BaseBdev1", 00:14:24.530 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.530 "is_configured": false, 00:14:24.530 "data_offset": 0, 00:14:24.530 "data_size": 0 00:14:24.530 }, 00:14:24.530 { 00:14:24.530 "name": "BaseBdev2", 00:14:24.530 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:24.530 "is_configured": true, 00:14:24.530 "data_offset": 2048, 00:14:24.530 "data_size": 63488 00:14:24.530 }, 00:14:24.530 { 00:14:24.530 "name": "BaseBdev3", 00:14:24.530 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:24.530 "is_configured": true, 00:14:24.530 "data_offset": 2048, 00:14:24.530 "data_size": 63488 00:14:24.530 }, 00:14:24.530 { 00:14:24.530 "name": "BaseBdev4", 00:14:24.530 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:24.530 "is_configured": true, 00:14:24.530 "data_offset": 2048, 00:14:24.530 "data_size": 63488 00:14:24.530 } 00:14:24.530 ] 00:14:24.530 }' 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.530 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.102 [2024-09-30 14:13:29.521926] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.102 "name": "Existed_Raid", 00:14:25.102 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:25.102 "strip_size_kb": 64, 00:14:25.102 "state": "configuring", 00:14:25.102 "raid_level": "raid5f", 00:14:25.102 "superblock": true, 00:14:25.102 "num_base_bdevs": 4, 00:14:25.102 "num_base_bdevs_discovered": 2, 00:14:25.102 "num_base_bdevs_operational": 4, 00:14:25.102 "base_bdevs_list": [ 00:14:25.102 { 00:14:25.102 "name": "BaseBdev1", 00:14:25.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:25.102 "is_configured": false, 00:14:25.102 "data_offset": 0, 00:14:25.102 "data_size": 0 00:14:25.102 }, 00:14:25.102 { 00:14:25.102 "name": null, 00:14:25.102 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:25.102 "is_configured": false, 00:14:25.102 "data_offset": 0, 00:14:25.102 "data_size": 63488 00:14:25.102 }, 00:14:25.102 { 00:14:25.102 "name": "BaseBdev3", 00:14:25.102 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:25.102 "is_configured": true, 00:14:25.102 "data_offset": 2048, 00:14:25.102 "data_size": 63488 00:14:25.102 }, 00:14:25.102 { 00:14:25.102 "name": "BaseBdev4", 00:14:25.102 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:25.102 "is_configured": true, 00:14:25.102 "data_offset": 2048, 00:14:25.102 "data_size": 63488 00:14:25.102 } 00:14:25.102 ] 00:14:25.102 }' 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.102 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:25.362 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.363 14:13:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.363 [2024-09-30 14:13:30.011863] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:25.363 BaseBdev1 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.363 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.623 [ 00:14:25.623 { 00:14:25.623 "name": "BaseBdev1", 00:14:25.623 "aliases": [ 00:14:25.623 "8e503cec-5462-444f-aa37-85b28eb59ba8" 00:14:25.623 ], 00:14:25.623 "product_name": "Malloc disk", 00:14:25.623 "block_size": 512, 00:14:25.623 "num_blocks": 65536, 00:14:25.623 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:25.623 "assigned_rate_limits": { 00:14:25.623 "rw_ios_per_sec": 0, 00:14:25.623 "rw_mbytes_per_sec": 0, 00:14:25.623 "r_mbytes_per_sec": 0, 00:14:25.623 "w_mbytes_per_sec": 0 00:14:25.623 }, 00:14:25.623 "claimed": true, 00:14:25.623 "claim_type": "exclusive_write", 00:14:25.623 "zoned": false, 00:14:25.623 "supported_io_types": { 00:14:25.623 "read": true, 00:14:25.623 "write": true, 00:14:25.623 "unmap": true, 00:14:25.623 "flush": true, 00:14:25.623 "reset": true, 00:14:25.623 "nvme_admin": false, 00:14:25.623 "nvme_io": false, 00:14:25.623 "nvme_io_md": false, 00:14:25.623 "write_zeroes": true, 00:14:25.623 "zcopy": true, 00:14:25.623 "get_zone_info": false, 00:14:25.623 "zone_management": false, 00:14:25.623 "zone_append": false, 00:14:25.623 "compare": false, 00:14:25.623 "compare_and_write": false, 00:14:25.623 "abort": true, 00:14:25.623 "seek_hole": false, 00:14:25.623 "seek_data": false, 00:14:25.623 "copy": true, 00:14:25.623 "nvme_iov_md": false 00:14:25.623 }, 00:14:25.623 "memory_domains": [ 00:14:25.623 { 00:14:25.623 "dma_device_id": "system", 00:14:25.623 "dma_device_type": 1 00:14:25.623 }, 00:14:25.623 { 00:14:25.623 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:25.623 "dma_device_type": 2 00:14:25.623 } 00:14:25.623 ], 00:14:25.623 "driver_specific": {} 00:14:25.623 } 00:14:25.623 ] 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.623 "name": "Existed_Raid", 00:14:25.623 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:25.623 "strip_size_kb": 64, 00:14:25.623 "state": "configuring", 00:14:25.623 "raid_level": "raid5f", 00:14:25.623 "superblock": true, 00:14:25.623 "num_base_bdevs": 4, 00:14:25.623 "num_base_bdevs_discovered": 3, 00:14:25.623 "num_base_bdevs_operational": 4, 00:14:25.623 "base_bdevs_list": [ 00:14:25.623 { 00:14:25.623 "name": "BaseBdev1", 00:14:25.623 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:25.623 "is_configured": true, 00:14:25.623 "data_offset": 2048, 00:14:25.623 "data_size": 63488 00:14:25.623 }, 00:14:25.623 { 00:14:25.623 "name": null, 00:14:25.623 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:25.623 "is_configured": false, 00:14:25.623 "data_offset": 0, 00:14:25.623 "data_size": 63488 00:14:25.623 }, 00:14:25.623 { 00:14:25.623 "name": "BaseBdev3", 00:14:25.623 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:25.623 "is_configured": true, 00:14:25.623 "data_offset": 2048, 00:14:25.623 "data_size": 63488 00:14:25.623 }, 00:14:25.623 { 00:14:25.623 "name": "BaseBdev4", 00:14:25.623 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:25.623 "is_configured": true, 00:14:25.623 "data_offset": 2048, 00:14:25.623 "data_size": 63488 00:14:25.623 } 00:14:25.623 ] 00:14:25.623 }' 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.623 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.883 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.883 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:25.883 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.883 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.883 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.143 [2024-09-30 14:13:30.555067] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.143 "name": "Existed_Raid", 00:14:26.143 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:26.143 "strip_size_kb": 64, 00:14:26.143 "state": "configuring", 00:14:26.143 "raid_level": "raid5f", 00:14:26.143 "superblock": true, 00:14:26.143 "num_base_bdevs": 4, 00:14:26.143 "num_base_bdevs_discovered": 2, 00:14:26.143 "num_base_bdevs_operational": 4, 00:14:26.143 "base_bdevs_list": [ 00:14:26.143 { 00:14:26.143 "name": "BaseBdev1", 00:14:26.143 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:26.143 "is_configured": true, 00:14:26.143 "data_offset": 2048, 00:14:26.143 "data_size": 63488 00:14:26.143 }, 00:14:26.143 { 00:14:26.143 "name": null, 00:14:26.143 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:26.143 "is_configured": false, 00:14:26.143 "data_offset": 0, 00:14:26.143 "data_size": 63488 00:14:26.143 }, 00:14:26.143 { 00:14:26.143 "name": null, 00:14:26.143 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:26.143 "is_configured": false, 00:14:26.143 "data_offset": 0, 00:14:26.143 "data_size": 63488 00:14:26.143 }, 00:14:26.143 { 00:14:26.143 "name": "BaseBdev4", 00:14:26.143 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:26.143 "is_configured": true, 00:14:26.143 "data_offset": 2048, 00:14:26.143 "data_size": 63488 00:14:26.143 } 00:14:26.143 ] 00:14:26.143 }' 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.143 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.403 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.403 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.403 14:13:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.403 14:13:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.403 [2024-09-30 14:13:31.042221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.403 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:26.663 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.663 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.663 "name": "Existed_Raid", 00:14:26.663 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:26.663 "strip_size_kb": 64, 00:14:26.663 "state": "configuring", 00:14:26.663 "raid_level": "raid5f", 00:14:26.663 "superblock": true, 00:14:26.663 "num_base_bdevs": 4, 00:14:26.663 "num_base_bdevs_discovered": 3, 00:14:26.663 "num_base_bdevs_operational": 4, 00:14:26.663 "base_bdevs_list": [ 00:14:26.663 { 00:14:26.663 "name": "BaseBdev1", 00:14:26.663 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:26.663 "is_configured": true, 00:14:26.663 "data_offset": 2048, 00:14:26.663 "data_size": 63488 00:14:26.663 }, 00:14:26.663 { 00:14:26.663 "name": null, 00:14:26.663 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:26.663 "is_configured": false, 00:14:26.663 "data_offset": 0, 00:14:26.663 "data_size": 63488 00:14:26.663 }, 00:14:26.663 { 00:14:26.663 "name": "BaseBdev3", 00:14:26.663 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:26.663 "is_configured": true, 00:14:26.663 "data_offset": 2048, 00:14:26.663 "data_size": 63488 00:14:26.663 }, 00:14:26.663 { 00:14:26.663 "name": "BaseBdev4", 00:14:26.663 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:26.663 "is_configured": true, 00:14:26.663 "data_offset": 2048, 00:14:26.663 "data_size": 63488 00:14:26.663 } 00:14:26.663 ] 00:14:26.663 }' 00:14:26.663 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.663 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.923 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.923 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.924 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.924 [2024-09-30 14:13:31.569375] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.184 "name": "Existed_Raid", 00:14:27.184 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:27.184 "strip_size_kb": 64, 00:14:27.184 "state": "configuring", 00:14:27.184 "raid_level": "raid5f", 00:14:27.184 "superblock": true, 00:14:27.184 "num_base_bdevs": 4, 00:14:27.184 "num_base_bdevs_discovered": 2, 00:14:27.184 "num_base_bdevs_operational": 4, 00:14:27.184 "base_bdevs_list": [ 00:14:27.184 { 00:14:27.184 "name": null, 00:14:27.184 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:27.184 "is_configured": false, 00:14:27.184 "data_offset": 0, 00:14:27.184 "data_size": 63488 00:14:27.184 }, 00:14:27.184 { 00:14:27.184 "name": null, 00:14:27.184 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:27.184 "is_configured": false, 00:14:27.184 "data_offset": 0, 00:14:27.184 "data_size": 63488 00:14:27.184 }, 00:14:27.184 { 00:14:27.184 "name": "BaseBdev3", 00:14:27.184 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:27.184 "is_configured": true, 00:14:27.184 "data_offset": 2048, 00:14:27.184 "data_size": 63488 00:14:27.184 }, 00:14:27.184 { 00:14:27.184 "name": "BaseBdev4", 00:14:27.184 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:27.184 "is_configured": true, 00:14:27.184 "data_offset": 2048, 00:14:27.184 "data_size": 63488 00:14:27.184 } 00:14:27.184 ] 00:14:27.184 }' 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.184 14:13:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.445 [2024-09-30 14:13:32.079149] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.445 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.705 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.705 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.705 "name": "Existed_Raid", 00:14:27.705 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:27.705 "strip_size_kb": 64, 00:14:27.705 "state": "configuring", 00:14:27.705 "raid_level": "raid5f", 00:14:27.705 "superblock": true, 00:14:27.705 "num_base_bdevs": 4, 00:14:27.705 "num_base_bdevs_discovered": 3, 00:14:27.705 "num_base_bdevs_operational": 4, 00:14:27.705 "base_bdevs_list": [ 00:14:27.705 { 00:14:27.705 "name": null, 00:14:27.705 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:27.705 "is_configured": false, 00:14:27.705 "data_offset": 0, 00:14:27.705 "data_size": 63488 00:14:27.705 }, 00:14:27.705 { 00:14:27.705 "name": "BaseBdev2", 00:14:27.705 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:27.705 "is_configured": true, 00:14:27.705 "data_offset": 2048, 00:14:27.705 "data_size": 63488 00:14:27.705 }, 00:14:27.705 { 00:14:27.705 "name": "BaseBdev3", 00:14:27.705 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:27.705 "is_configured": true, 00:14:27.705 "data_offset": 2048, 00:14:27.705 "data_size": 63488 00:14:27.705 }, 00:14:27.705 { 00:14:27.705 "name": "BaseBdev4", 00:14:27.705 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:27.705 "is_configured": true, 00:14:27.705 "data_offset": 2048, 00:14:27.705 "data_size": 63488 00:14:27.705 } 00:14:27.705 ] 00:14:27.705 }' 00:14:27.705 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.705 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8e503cec-5462-444f-aa37-85b28eb59ba8 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.965 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.226 [2024-09-30 14:13:32.629204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:28.226 [2024-09-30 14:13:32.629428] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:14:28.226 [2024-09-30 14:13:32.629448] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:28.226 [2024-09-30 14:13:32.629718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006490 00:14:28.226 NewBaseBdev 00:14:28.226 [2024-09-30 14:13:32.630144] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:14:28.226 [2024-09-30 14:13:32.630156] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:14:28.226 [2024-09-30 14:13:32.630250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.226 [ 00:14:28.226 { 00:14:28.226 "name": "NewBaseBdev", 00:14:28.226 "aliases": [ 00:14:28.226 "8e503cec-5462-444f-aa37-85b28eb59ba8" 00:14:28.226 ], 00:14:28.226 "product_name": "Malloc disk", 00:14:28.226 "block_size": 512, 00:14:28.226 "num_blocks": 65536, 00:14:28.226 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:28.226 "assigned_rate_limits": { 00:14:28.226 "rw_ios_per_sec": 0, 00:14:28.226 "rw_mbytes_per_sec": 0, 00:14:28.226 "r_mbytes_per_sec": 0, 00:14:28.226 "w_mbytes_per_sec": 0 00:14:28.226 }, 00:14:28.226 "claimed": true, 00:14:28.226 "claim_type": "exclusive_write", 00:14:28.226 "zoned": false, 00:14:28.226 "supported_io_types": { 00:14:28.226 "read": true, 00:14:28.226 "write": true, 00:14:28.226 "unmap": true, 00:14:28.226 "flush": true, 00:14:28.226 "reset": true, 00:14:28.226 "nvme_admin": false, 00:14:28.226 "nvme_io": false, 00:14:28.226 "nvme_io_md": false, 00:14:28.226 "write_zeroes": true, 00:14:28.226 "zcopy": true, 00:14:28.226 "get_zone_info": false, 00:14:28.226 "zone_management": false, 00:14:28.226 "zone_append": false, 00:14:28.226 "compare": false, 00:14:28.226 "compare_and_write": false, 00:14:28.226 "abort": true, 00:14:28.226 "seek_hole": false, 00:14:28.226 "seek_data": false, 00:14:28.226 "copy": true, 00:14:28.226 "nvme_iov_md": false 00:14:28.226 }, 00:14:28.226 "memory_domains": [ 00:14:28.226 { 00:14:28.226 "dma_device_id": "system", 00:14:28.226 "dma_device_type": 1 00:14:28.226 }, 00:14:28.226 { 00:14:28.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:28.226 "dma_device_type": 2 00:14:28.226 } 00:14:28.226 ], 00:14:28.226 "driver_specific": {} 00:14:28.226 } 00:14:28.226 ] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.226 "name": "Existed_Raid", 00:14:28.226 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:28.226 "strip_size_kb": 64, 00:14:28.226 "state": "online", 00:14:28.226 "raid_level": "raid5f", 00:14:28.226 "superblock": true, 00:14:28.226 "num_base_bdevs": 4, 00:14:28.226 "num_base_bdevs_discovered": 4, 00:14:28.226 "num_base_bdevs_operational": 4, 00:14:28.226 "base_bdevs_list": [ 00:14:28.226 { 00:14:28.226 "name": "NewBaseBdev", 00:14:28.226 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:28.226 "is_configured": true, 00:14:28.226 "data_offset": 2048, 00:14:28.226 "data_size": 63488 00:14:28.226 }, 00:14:28.226 { 00:14:28.226 "name": "BaseBdev2", 00:14:28.226 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:28.226 "is_configured": true, 00:14:28.226 "data_offset": 2048, 00:14:28.226 "data_size": 63488 00:14:28.226 }, 00:14:28.226 { 00:14:28.226 "name": "BaseBdev3", 00:14:28.226 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:28.226 "is_configured": true, 00:14:28.226 "data_offset": 2048, 00:14:28.226 "data_size": 63488 00:14:28.226 }, 00:14:28.226 { 00:14:28.226 "name": "BaseBdev4", 00:14:28.226 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:28.226 "is_configured": true, 00:14:28.226 "data_offset": 2048, 00:14:28.226 "data_size": 63488 00:14:28.226 } 00:14:28.226 ] 00:14:28.226 }' 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.226 14:13:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.486 [2024-09-30 14:13:33.100623] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.486 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:28.486 "name": "Existed_Raid", 00:14:28.486 "aliases": [ 00:14:28.486 "09ee10f7-3c24-4b84-9468-4906e5463231" 00:14:28.486 ], 00:14:28.486 "product_name": "Raid Volume", 00:14:28.486 "block_size": 512, 00:14:28.486 "num_blocks": 190464, 00:14:28.486 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:28.486 "assigned_rate_limits": { 00:14:28.486 "rw_ios_per_sec": 0, 00:14:28.486 "rw_mbytes_per_sec": 0, 00:14:28.486 "r_mbytes_per_sec": 0, 00:14:28.486 "w_mbytes_per_sec": 0 00:14:28.486 }, 00:14:28.486 "claimed": false, 00:14:28.486 "zoned": false, 00:14:28.486 "supported_io_types": { 00:14:28.486 "read": true, 00:14:28.486 "write": true, 00:14:28.486 "unmap": false, 00:14:28.486 "flush": false, 00:14:28.486 "reset": true, 00:14:28.486 "nvme_admin": false, 00:14:28.486 "nvme_io": false, 00:14:28.486 "nvme_io_md": false, 00:14:28.486 "write_zeroes": true, 00:14:28.486 "zcopy": false, 00:14:28.486 "get_zone_info": false, 00:14:28.486 "zone_management": false, 00:14:28.486 "zone_append": false, 00:14:28.486 "compare": false, 00:14:28.486 "compare_and_write": false, 00:14:28.486 "abort": false, 00:14:28.486 "seek_hole": false, 00:14:28.486 "seek_data": false, 00:14:28.486 "copy": false, 00:14:28.486 "nvme_iov_md": false 00:14:28.486 }, 00:14:28.486 "driver_specific": { 00:14:28.486 "raid": { 00:14:28.486 "uuid": "09ee10f7-3c24-4b84-9468-4906e5463231", 00:14:28.486 "strip_size_kb": 64, 00:14:28.486 "state": "online", 00:14:28.486 "raid_level": "raid5f", 00:14:28.486 "superblock": true, 00:14:28.486 "num_base_bdevs": 4, 00:14:28.486 "num_base_bdevs_discovered": 4, 00:14:28.486 "num_base_bdevs_operational": 4, 00:14:28.486 "base_bdevs_list": [ 00:14:28.486 { 00:14:28.486 "name": "NewBaseBdev", 00:14:28.486 "uuid": "8e503cec-5462-444f-aa37-85b28eb59ba8", 00:14:28.486 "is_configured": true, 00:14:28.486 "data_offset": 2048, 00:14:28.486 "data_size": 63488 00:14:28.486 }, 00:14:28.486 { 00:14:28.486 "name": "BaseBdev2", 00:14:28.486 "uuid": "639f25a8-a4b1-4263-8dbf-43592e64a60f", 00:14:28.486 "is_configured": true, 00:14:28.486 "data_offset": 2048, 00:14:28.486 "data_size": 63488 00:14:28.486 }, 00:14:28.487 { 00:14:28.487 "name": "BaseBdev3", 00:14:28.487 "uuid": "eb2a6e28-6d70-4515-895f-c6cefe6e441d", 00:14:28.487 "is_configured": true, 00:14:28.487 "data_offset": 2048, 00:14:28.487 "data_size": 63488 00:14:28.487 }, 00:14:28.487 { 00:14:28.487 "name": "BaseBdev4", 00:14:28.487 "uuid": "a276601c-fec4-4f2f-805a-7e218f02d601", 00:14:28.487 "is_configured": true, 00:14:28.487 "data_offset": 2048, 00:14:28.487 "data_size": 63488 00:14:28.487 } 00:14:28.487 ] 00:14:28.487 } 00:14:28.487 } 00:14:28.487 }' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:28.747 BaseBdev2 00:14:28.747 BaseBdev3 00:14:28.747 BaseBdev4' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.747 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.747 [2024-09-30 14:13:33.399926] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:28.747 [2024-09-30 14:13:33.399950] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:28.747 [2024-09-30 14:13:33.400012] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:28.747 [2024-09-30 14:13:33.400253] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:28.747 [2024-09-30 14:13:33.400266] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 94919 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 94919 ']' 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 94919 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94919 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:29.007 killing process with pid 94919 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94919' 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 94919 00:14:29.007 [2024-09-30 14:13:33.440922] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:29.007 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 94919 00:14:29.007 [2024-09-30 14:13:33.481345] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:29.267 14:13:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:29.267 ************************************ 00:14:29.267 END TEST raid5f_state_function_test_sb 00:14:29.267 ************************************ 00:14:29.267 00:14:29.267 real 0m9.572s 00:14:29.267 user 0m16.246s 00:14:29.267 sys 0m2.118s 00:14:29.267 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:29.268 14:13:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.268 14:13:33 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:14:29.268 14:13:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:29.268 14:13:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:29.268 14:13:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:29.268 ************************************ 00:14:29.268 START TEST raid5f_superblock_test 00:14:29.268 ************************************ 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 4 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=95567 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 95567 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 95567 ']' 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:29.268 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:29.268 14:13:33 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.268 [2024-09-30 14:13:33.905082] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:14:29.268 [2024-09-30 14:13:33.905282] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95567 ] 00:14:29.528 [2024-09-30 14:13:34.041399] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:14:29.528 [2024-09-30 14:13:34.071441] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:29.528 [2024-09-30 14:13:34.117880] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:29.528 [2024-09-30 14:13:34.159901] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:29.528 [2024-09-30 14:13:34.160007] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.099 malloc1 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.099 [2024-09-30 14:13:34.741986] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:30.099 [2024-09-30 14:13:34.742056] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.099 [2024-09-30 14:13:34.742085] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:30.099 [2024-09-30 14:13:34.742096] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.099 [2024-09-30 14:13:34.744138] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.099 [2024-09-30 14:13:34.744179] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:30.099 pt1 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.099 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 malloc2 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 [2024-09-30 14:13:34.784686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:30.360 [2024-09-30 14:13:34.784824] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.360 [2024-09-30 14:13:34.784877] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:30.360 [2024-09-30 14:13:34.784906] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.360 [2024-09-30 14:13:34.786878] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.360 [2024-09-30 14:13:34.786943] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:30.360 pt2 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 malloc3 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 [2024-09-30 14:13:34.816955] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:30.360 [2024-09-30 14:13:34.817048] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.360 [2024-09-30 14:13:34.817072] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:30.360 [2024-09-30 14:13:34.817080] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.360 [2024-09-30 14:13:34.819021] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.360 [2024-09-30 14:13:34.819065] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:30.360 pt3 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 malloc4 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 [2024-09-30 14:13:34.849235] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:30.360 [2024-09-30 14:13:34.849328] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.360 [2024-09-30 14:13:34.849363] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:30.360 [2024-09-30 14:13:34.849390] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.360 [2024-09-30 14:13:34.851384] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.360 [2024-09-30 14:13:34.851449] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:30.360 pt4 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.360 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.360 [2024-09-30 14:13:34.861256] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:30.360 [2024-09-30 14:13:34.863052] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:30.360 [2024-09-30 14:13:34.863152] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:30.360 [2024-09-30 14:13:34.863228] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:30.360 [2024-09-30 14:13:34.863444] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:14:30.360 [2024-09-30 14:13:34.863503] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:30.361 [2024-09-30 14:13:34.863796] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:30.361 [2024-09-30 14:13:34.864260] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:14:30.361 [2024-09-30 14:13:34.864306] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:14:30.361 [2024-09-30 14:13:34.864452] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:30.361 "name": "raid_bdev1", 00:14:30.361 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:30.361 "strip_size_kb": 64, 00:14:30.361 "state": "online", 00:14:30.361 "raid_level": "raid5f", 00:14:30.361 "superblock": true, 00:14:30.361 "num_base_bdevs": 4, 00:14:30.361 "num_base_bdevs_discovered": 4, 00:14:30.361 "num_base_bdevs_operational": 4, 00:14:30.361 "base_bdevs_list": [ 00:14:30.361 { 00:14:30.361 "name": "pt1", 00:14:30.361 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:30.361 "is_configured": true, 00:14:30.361 "data_offset": 2048, 00:14:30.361 "data_size": 63488 00:14:30.361 }, 00:14:30.361 { 00:14:30.361 "name": "pt2", 00:14:30.361 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:30.361 "is_configured": true, 00:14:30.361 "data_offset": 2048, 00:14:30.361 "data_size": 63488 00:14:30.361 }, 00:14:30.361 { 00:14:30.361 "name": "pt3", 00:14:30.361 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:30.361 "is_configured": true, 00:14:30.361 "data_offset": 2048, 00:14:30.361 "data_size": 63488 00:14:30.361 }, 00:14:30.361 { 00:14:30.361 "name": "pt4", 00:14:30.361 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:30.361 "is_configured": true, 00:14:30.361 "data_offset": 2048, 00:14:30.361 "data_size": 63488 00:14:30.361 } 00:14:30.361 ] 00:14:30.361 }' 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:30.361 14:13:34 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.931 [2024-09-30 14:13:35.369483] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:30.931 "name": "raid_bdev1", 00:14:30.931 "aliases": [ 00:14:30.931 "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1" 00:14:30.931 ], 00:14:30.931 "product_name": "Raid Volume", 00:14:30.931 "block_size": 512, 00:14:30.931 "num_blocks": 190464, 00:14:30.931 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:30.931 "assigned_rate_limits": { 00:14:30.931 "rw_ios_per_sec": 0, 00:14:30.931 "rw_mbytes_per_sec": 0, 00:14:30.931 "r_mbytes_per_sec": 0, 00:14:30.931 "w_mbytes_per_sec": 0 00:14:30.931 }, 00:14:30.931 "claimed": false, 00:14:30.931 "zoned": false, 00:14:30.931 "supported_io_types": { 00:14:30.931 "read": true, 00:14:30.931 "write": true, 00:14:30.931 "unmap": false, 00:14:30.931 "flush": false, 00:14:30.931 "reset": true, 00:14:30.931 "nvme_admin": false, 00:14:30.931 "nvme_io": false, 00:14:30.931 "nvme_io_md": false, 00:14:30.931 "write_zeroes": true, 00:14:30.931 "zcopy": false, 00:14:30.931 "get_zone_info": false, 00:14:30.931 "zone_management": false, 00:14:30.931 "zone_append": false, 00:14:30.931 "compare": false, 00:14:30.931 "compare_and_write": false, 00:14:30.931 "abort": false, 00:14:30.931 "seek_hole": false, 00:14:30.931 "seek_data": false, 00:14:30.931 "copy": false, 00:14:30.931 "nvme_iov_md": false 00:14:30.931 }, 00:14:30.931 "driver_specific": { 00:14:30.931 "raid": { 00:14:30.931 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:30.931 "strip_size_kb": 64, 00:14:30.931 "state": "online", 00:14:30.931 "raid_level": "raid5f", 00:14:30.931 "superblock": true, 00:14:30.931 "num_base_bdevs": 4, 00:14:30.931 "num_base_bdevs_discovered": 4, 00:14:30.931 "num_base_bdevs_operational": 4, 00:14:30.931 "base_bdevs_list": [ 00:14:30.931 { 00:14:30.931 "name": "pt1", 00:14:30.931 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:30.931 "is_configured": true, 00:14:30.931 "data_offset": 2048, 00:14:30.931 "data_size": 63488 00:14:30.931 }, 00:14:30.931 { 00:14:30.931 "name": "pt2", 00:14:30.931 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:30.931 "is_configured": true, 00:14:30.931 "data_offset": 2048, 00:14:30.931 "data_size": 63488 00:14:30.931 }, 00:14:30.931 { 00:14:30.931 "name": "pt3", 00:14:30.931 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:30.931 "is_configured": true, 00:14:30.931 "data_offset": 2048, 00:14:30.931 "data_size": 63488 00:14:30.931 }, 00:14:30.931 { 00:14:30.931 "name": "pt4", 00:14:30.931 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:30.931 "is_configured": true, 00:14:30.931 "data_offset": 2048, 00:14:30.931 "data_size": 63488 00:14:30.931 } 00:14:30.931 ] 00:14:30.931 } 00:14:30.931 } 00:14:30.931 }' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:30.931 pt2 00:14:30.931 pt3 00:14:30.931 pt4' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:30.931 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 [2024-09-30 14:13:35.680923] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 ']' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 [2024-09-30 14:13:35.720709] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:31.192 [2024-09-30 14:13:35.720733] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:31.192 [2024-09-30 14:13:35.720812] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:31.192 [2024-09-30 14:13:35.720888] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:31.192 [2024-09-30 14:13:35.720900] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.192 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:31.452 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.453 [2024-09-30 14:13:35.888437] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:31.453 [2024-09-30 14:13:35.890236] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:31.453 [2024-09-30 14:13:35.890316] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:31.453 [2024-09-30 14:13:35.890364] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:14:31.453 [2024-09-30 14:13:35.890428] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:31.453 [2024-09-30 14:13:35.890513] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:31.453 [2024-09-30 14:13:35.890562] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:31.453 [2024-09-30 14:13:35.890580] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:14:31.453 [2024-09-30 14:13:35.890592] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:31.453 [2024-09-30 14:13:35.890602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:14:31.453 request: 00:14:31.453 { 00:14:31.453 "name": "raid_bdev1", 00:14:31.453 "raid_level": "raid5f", 00:14:31.453 "base_bdevs": [ 00:14:31.453 "malloc1", 00:14:31.453 "malloc2", 00:14:31.453 "malloc3", 00:14:31.453 "malloc4" 00:14:31.453 ], 00:14:31.453 "strip_size_kb": 64, 00:14:31.453 "superblock": false, 00:14:31.453 "method": "bdev_raid_create", 00:14:31.453 "req_id": 1 00:14:31.453 } 00:14:31.453 Got JSON-RPC error response 00:14:31.453 response: 00:14:31.453 { 00:14:31.453 "code": -17, 00:14:31.453 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:31.453 } 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.453 [2024-09-30 14:13:35.952302] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:31.453 [2024-09-30 14:13:35.952361] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:31.453 [2024-09-30 14:13:35.952392] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:31.453 [2024-09-30 14:13:35.952402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:31.453 [2024-09-30 14:13:35.954367] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:31.453 [2024-09-30 14:13:35.954405] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:31.453 [2024-09-30 14:13:35.954463] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:31.453 [2024-09-30 14:13:35.954528] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:31.453 pt1 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.453 14:13:35 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.453 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.453 "name": "raid_bdev1", 00:14:31.453 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:31.453 "strip_size_kb": 64, 00:14:31.453 "state": "configuring", 00:14:31.453 "raid_level": "raid5f", 00:14:31.453 "superblock": true, 00:14:31.453 "num_base_bdevs": 4, 00:14:31.453 "num_base_bdevs_discovered": 1, 00:14:31.453 "num_base_bdevs_operational": 4, 00:14:31.453 "base_bdevs_list": [ 00:14:31.453 { 00:14:31.453 "name": "pt1", 00:14:31.453 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:31.453 "is_configured": true, 00:14:31.453 "data_offset": 2048, 00:14:31.453 "data_size": 63488 00:14:31.453 }, 00:14:31.453 { 00:14:31.453 "name": null, 00:14:31.453 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:31.453 "is_configured": false, 00:14:31.453 "data_offset": 2048, 00:14:31.453 "data_size": 63488 00:14:31.453 }, 00:14:31.453 { 00:14:31.453 "name": null, 00:14:31.453 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:31.453 "is_configured": false, 00:14:31.453 "data_offset": 2048, 00:14:31.453 "data_size": 63488 00:14:31.453 }, 00:14:31.453 { 00:14:31.453 "name": null, 00:14:31.453 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:31.453 "is_configured": false, 00:14:31.453 "data_offset": 2048, 00:14:31.453 "data_size": 63488 00:14:31.453 } 00:14:31.453 ] 00:14:31.453 }' 00:14:31.453 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.453 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.022 [2024-09-30 14:13:36.399521] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:32.022 [2024-09-30 14:13:36.399617] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.022 [2024-09-30 14:13:36.399651] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:32.022 [2024-09-30 14:13:36.399681] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.022 [2024-09-30 14:13:36.400021] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.022 [2024-09-30 14:13:36.400074] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:32.022 [2024-09-30 14:13:36.400154] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:32.022 [2024-09-30 14:13:36.400203] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:32.022 pt2 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.022 [2024-09-30 14:13:36.411521] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.022 "name": "raid_bdev1", 00:14:32.022 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:32.022 "strip_size_kb": 64, 00:14:32.022 "state": "configuring", 00:14:32.022 "raid_level": "raid5f", 00:14:32.022 "superblock": true, 00:14:32.022 "num_base_bdevs": 4, 00:14:32.022 "num_base_bdevs_discovered": 1, 00:14:32.022 "num_base_bdevs_operational": 4, 00:14:32.022 "base_bdevs_list": [ 00:14:32.022 { 00:14:32.022 "name": "pt1", 00:14:32.022 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:32.022 "is_configured": true, 00:14:32.022 "data_offset": 2048, 00:14:32.022 "data_size": 63488 00:14:32.022 }, 00:14:32.022 { 00:14:32.022 "name": null, 00:14:32.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.022 "is_configured": false, 00:14:32.022 "data_offset": 0, 00:14:32.022 "data_size": 63488 00:14:32.022 }, 00:14:32.022 { 00:14:32.022 "name": null, 00:14:32.022 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.022 "is_configured": false, 00:14:32.022 "data_offset": 2048, 00:14:32.022 "data_size": 63488 00:14:32.022 }, 00:14:32.022 { 00:14:32.022 "name": null, 00:14:32.022 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:32.022 "is_configured": false, 00:14:32.022 "data_offset": 2048, 00:14:32.022 "data_size": 63488 00:14:32.022 } 00:14:32.022 ] 00:14:32.022 }' 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.022 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.283 [2024-09-30 14:13:36.854800] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:32.283 [2024-09-30 14:13:36.854895] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.283 [2024-09-30 14:13:36.854917] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:32.283 [2024-09-30 14:13:36.854925] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.283 [2024-09-30 14:13:36.855256] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.283 [2024-09-30 14:13:36.855278] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:32.283 [2024-09-30 14:13:36.855364] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:32.283 [2024-09-30 14:13:36.855390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:32.283 pt2 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.283 [2024-09-30 14:13:36.866748] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:32.283 [2024-09-30 14:13:36.866791] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.283 [2024-09-30 14:13:36.866822] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:32.283 [2024-09-30 14:13:36.866830] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.283 [2024-09-30 14:13:36.867128] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.283 [2024-09-30 14:13:36.867148] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:32.283 [2024-09-30 14:13:36.867202] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:32.283 [2024-09-30 14:13:36.867224] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:32.283 pt3 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.283 [2024-09-30 14:13:36.878722] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:32.283 [2024-09-30 14:13:36.878767] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.283 [2024-09-30 14:13:36.878800] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:32.283 [2024-09-30 14:13:36.878807] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.283 [2024-09-30 14:13:36.879090] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.283 [2024-09-30 14:13:36.879103] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:32.283 [2024-09-30 14:13:36.879151] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:32.283 [2024-09-30 14:13:36.879173] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:32.283 [2024-09-30 14:13:36.879275] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:14:32.283 [2024-09-30 14:13:36.879283] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:32.283 [2024-09-30 14:13:36.879524] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:14:32.283 [2024-09-30 14:13:36.879998] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:14:32.283 [2024-09-30 14:13:36.880012] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:14:32.283 [2024-09-30 14:13:36.880102] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:32.283 pt4 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.283 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.542 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.542 "name": "raid_bdev1", 00:14:32.542 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:32.542 "strip_size_kb": 64, 00:14:32.542 "state": "online", 00:14:32.542 "raid_level": "raid5f", 00:14:32.542 "superblock": true, 00:14:32.542 "num_base_bdevs": 4, 00:14:32.542 "num_base_bdevs_discovered": 4, 00:14:32.542 "num_base_bdevs_operational": 4, 00:14:32.542 "base_bdevs_list": [ 00:14:32.542 { 00:14:32.542 "name": "pt1", 00:14:32.542 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:32.542 "is_configured": true, 00:14:32.542 "data_offset": 2048, 00:14:32.542 "data_size": 63488 00:14:32.542 }, 00:14:32.542 { 00:14:32.542 "name": "pt2", 00:14:32.542 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.542 "is_configured": true, 00:14:32.542 "data_offset": 2048, 00:14:32.542 "data_size": 63488 00:14:32.542 }, 00:14:32.542 { 00:14:32.542 "name": "pt3", 00:14:32.542 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.542 "is_configured": true, 00:14:32.542 "data_offset": 2048, 00:14:32.542 "data_size": 63488 00:14:32.542 }, 00:14:32.542 { 00:14:32.542 "name": "pt4", 00:14:32.542 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:32.542 "is_configured": true, 00:14:32.542 "data_offset": 2048, 00:14:32.542 "data_size": 63488 00:14:32.542 } 00:14:32.542 ] 00:14:32.542 }' 00:14:32.542 14:13:36 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.542 14:13:36 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.802 [2024-09-30 14:13:37.326135] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.802 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:32.802 "name": "raid_bdev1", 00:14:32.802 "aliases": [ 00:14:32.802 "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1" 00:14:32.802 ], 00:14:32.802 "product_name": "Raid Volume", 00:14:32.802 "block_size": 512, 00:14:32.802 "num_blocks": 190464, 00:14:32.802 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:32.802 "assigned_rate_limits": { 00:14:32.802 "rw_ios_per_sec": 0, 00:14:32.802 "rw_mbytes_per_sec": 0, 00:14:32.802 "r_mbytes_per_sec": 0, 00:14:32.802 "w_mbytes_per_sec": 0 00:14:32.802 }, 00:14:32.802 "claimed": false, 00:14:32.802 "zoned": false, 00:14:32.802 "supported_io_types": { 00:14:32.802 "read": true, 00:14:32.802 "write": true, 00:14:32.802 "unmap": false, 00:14:32.802 "flush": false, 00:14:32.802 "reset": true, 00:14:32.802 "nvme_admin": false, 00:14:32.802 "nvme_io": false, 00:14:32.802 "nvme_io_md": false, 00:14:32.802 "write_zeroes": true, 00:14:32.802 "zcopy": false, 00:14:32.802 "get_zone_info": false, 00:14:32.802 "zone_management": false, 00:14:32.802 "zone_append": false, 00:14:32.802 "compare": false, 00:14:32.802 "compare_and_write": false, 00:14:32.802 "abort": false, 00:14:32.802 "seek_hole": false, 00:14:32.802 "seek_data": false, 00:14:32.802 "copy": false, 00:14:32.802 "nvme_iov_md": false 00:14:32.802 }, 00:14:32.802 "driver_specific": { 00:14:32.802 "raid": { 00:14:32.802 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:32.802 "strip_size_kb": 64, 00:14:32.802 "state": "online", 00:14:32.802 "raid_level": "raid5f", 00:14:32.802 "superblock": true, 00:14:32.802 "num_base_bdevs": 4, 00:14:32.802 "num_base_bdevs_discovered": 4, 00:14:32.802 "num_base_bdevs_operational": 4, 00:14:32.802 "base_bdevs_list": [ 00:14:32.802 { 00:14:32.802 "name": "pt1", 00:14:32.802 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:32.802 "is_configured": true, 00:14:32.802 "data_offset": 2048, 00:14:32.802 "data_size": 63488 00:14:32.802 }, 00:14:32.802 { 00:14:32.802 "name": "pt2", 00:14:32.802 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.802 "is_configured": true, 00:14:32.802 "data_offset": 2048, 00:14:32.802 "data_size": 63488 00:14:32.802 }, 00:14:32.802 { 00:14:32.802 "name": "pt3", 00:14:32.802 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.802 "is_configured": true, 00:14:32.803 "data_offset": 2048, 00:14:32.803 "data_size": 63488 00:14:32.803 }, 00:14:32.803 { 00:14:32.803 "name": "pt4", 00:14:32.803 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:32.803 "is_configured": true, 00:14:32.803 "data_offset": 2048, 00:14:32.803 "data_size": 63488 00:14:32.803 } 00:14:32.803 ] 00:14:32.803 } 00:14:32.803 } 00:14:32.803 }' 00:14:32.803 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:32.803 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:32.803 pt2 00:14:32.803 pt3 00:14:32.803 pt4' 00:14:32.803 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:33.063 [2024-09-30 14:13:37.657538] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 '!=' eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 ']' 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.063 [2024-09-30 14:13:37.705355] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.063 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.323 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.323 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.323 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.323 "name": "raid_bdev1", 00:14:33.323 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:33.323 "strip_size_kb": 64, 00:14:33.323 "state": "online", 00:14:33.323 "raid_level": "raid5f", 00:14:33.323 "superblock": true, 00:14:33.323 "num_base_bdevs": 4, 00:14:33.323 "num_base_bdevs_discovered": 3, 00:14:33.323 "num_base_bdevs_operational": 3, 00:14:33.323 "base_bdevs_list": [ 00:14:33.323 { 00:14:33.323 "name": null, 00:14:33.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.323 "is_configured": false, 00:14:33.323 "data_offset": 0, 00:14:33.323 "data_size": 63488 00:14:33.323 }, 00:14:33.323 { 00:14:33.323 "name": "pt2", 00:14:33.323 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:33.323 "is_configured": true, 00:14:33.323 "data_offset": 2048, 00:14:33.323 "data_size": 63488 00:14:33.323 }, 00:14:33.323 { 00:14:33.323 "name": "pt3", 00:14:33.323 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:33.323 "is_configured": true, 00:14:33.323 "data_offset": 2048, 00:14:33.323 "data_size": 63488 00:14:33.323 }, 00:14:33.323 { 00:14:33.323 "name": "pt4", 00:14:33.323 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:33.323 "is_configured": true, 00:14:33.323 "data_offset": 2048, 00:14:33.323 "data_size": 63488 00:14:33.323 } 00:14:33.323 ] 00:14:33.323 }' 00:14:33.323 14:13:37 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.323 14:13:37 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 [2024-09-30 14:13:38.104619] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:33.583 [2024-09-30 14:13:38.104697] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:33.583 [2024-09-30 14:13:38.104779] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:33.583 [2024-09-30 14:13:38.104857] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:33.583 [2024-09-30 14:13:38.104923] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 [2024-09-30 14:13:38.184533] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:33.583 [2024-09-30 14:13:38.184577] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:33.583 [2024-09-30 14:13:38.184594] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:33.583 [2024-09-30 14:13:38.184602] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:33.583 [2024-09-30 14:13:38.186783] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:33.583 [2024-09-30 14:13:38.186817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:33.583 [2024-09-30 14:13:38.186879] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:33.583 [2024-09-30 14:13:38.186908] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:33.583 pt2 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.583 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.843 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.843 "name": "raid_bdev1", 00:14:33.843 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:33.843 "strip_size_kb": 64, 00:14:33.843 "state": "configuring", 00:14:33.843 "raid_level": "raid5f", 00:14:33.843 "superblock": true, 00:14:33.843 "num_base_bdevs": 4, 00:14:33.843 "num_base_bdevs_discovered": 1, 00:14:33.843 "num_base_bdevs_operational": 3, 00:14:33.843 "base_bdevs_list": [ 00:14:33.843 { 00:14:33.843 "name": null, 00:14:33.843 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.843 "is_configured": false, 00:14:33.843 "data_offset": 2048, 00:14:33.843 "data_size": 63488 00:14:33.843 }, 00:14:33.843 { 00:14:33.843 "name": "pt2", 00:14:33.843 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:33.843 "is_configured": true, 00:14:33.843 "data_offset": 2048, 00:14:33.843 "data_size": 63488 00:14:33.843 }, 00:14:33.843 { 00:14:33.843 "name": null, 00:14:33.843 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:33.843 "is_configured": false, 00:14:33.843 "data_offset": 2048, 00:14:33.843 "data_size": 63488 00:14:33.843 }, 00:14:33.843 { 00:14:33.843 "name": null, 00:14:33.843 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:33.843 "is_configured": false, 00:14:33.843 "data_offset": 2048, 00:14:33.843 "data_size": 63488 00:14:33.843 } 00:14:33.843 ] 00:14:33.843 }' 00:14:33.843 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.843 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.103 [2024-09-30 14:13:38.623787] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:34.103 [2024-09-30 14:13:38.623883] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:34.103 [2024-09-30 14:13:38.623921] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:34.103 [2024-09-30 14:13:38.623948] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:34.103 [2024-09-30 14:13:38.624321] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:34.103 [2024-09-30 14:13:38.624375] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:34.103 [2024-09-30 14:13:38.624462] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:34.103 [2024-09-30 14:13:38.624539] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:34.103 pt3 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.103 "name": "raid_bdev1", 00:14:34.103 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:34.103 "strip_size_kb": 64, 00:14:34.103 "state": "configuring", 00:14:34.103 "raid_level": "raid5f", 00:14:34.103 "superblock": true, 00:14:34.103 "num_base_bdevs": 4, 00:14:34.103 "num_base_bdevs_discovered": 2, 00:14:34.103 "num_base_bdevs_operational": 3, 00:14:34.103 "base_bdevs_list": [ 00:14:34.103 { 00:14:34.103 "name": null, 00:14:34.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.103 "is_configured": false, 00:14:34.103 "data_offset": 2048, 00:14:34.103 "data_size": 63488 00:14:34.103 }, 00:14:34.103 { 00:14:34.103 "name": "pt2", 00:14:34.103 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:34.103 "is_configured": true, 00:14:34.103 "data_offset": 2048, 00:14:34.103 "data_size": 63488 00:14:34.103 }, 00:14:34.103 { 00:14:34.103 "name": "pt3", 00:14:34.103 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:34.103 "is_configured": true, 00:14:34.103 "data_offset": 2048, 00:14:34.103 "data_size": 63488 00:14:34.103 }, 00:14:34.103 { 00:14:34.103 "name": null, 00:14:34.103 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:34.103 "is_configured": false, 00:14:34.103 "data_offset": 2048, 00:14:34.103 "data_size": 63488 00:14:34.103 } 00:14:34.103 ] 00:14:34.103 }' 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.103 14:13:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.672 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.672 [2024-09-30 14:13:39.055080] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:34.672 [2024-09-30 14:13:39.055130] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:34.673 [2024-09-30 14:13:39.055168] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:34.673 [2024-09-30 14:13:39.055176] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:34.673 [2024-09-30 14:13:39.055538] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:34.673 [2024-09-30 14:13:39.055555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:34.673 [2024-09-30 14:13:39.055618] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:34.673 [2024-09-30 14:13:39.055638] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:34.673 [2024-09-30 14:13:39.055730] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:14:34.673 [2024-09-30 14:13:39.055738] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:34.673 [2024-09-30 14:13:39.055957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:14:34.673 [2024-09-30 14:13:39.056503] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:14:34.673 [2024-09-30 14:13:39.056523] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:14:34.673 [2024-09-30 14:13:39.056733] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:34.673 pt4 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.673 "name": "raid_bdev1", 00:14:34.673 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:34.673 "strip_size_kb": 64, 00:14:34.673 "state": "online", 00:14:34.673 "raid_level": "raid5f", 00:14:34.673 "superblock": true, 00:14:34.673 "num_base_bdevs": 4, 00:14:34.673 "num_base_bdevs_discovered": 3, 00:14:34.673 "num_base_bdevs_operational": 3, 00:14:34.673 "base_bdevs_list": [ 00:14:34.673 { 00:14:34.673 "name": null, 00:14:34.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.673 "is_configured": false, 00:14:34.673 "data_offset": 2048, 00:14:34.673 "data_size": 63488 00:14:34.673 }, 00:14:34.673 { 00:14:34.673 "name": "pt2", 00:14:34.673 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:34.673 "is_configured": true, 00:14:34.673 "data_offset": 2048, 00:14:34.673 "data_size": 63488 00:14:34.673 }, 00:14:34.673 { 00:14:34.673 "name": "pt3", 00:14:34.673 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:34.673 "is_configured": true, 00:14:34.673 "data_offset": 2048, 00:14:34.673 "data_size": 63488 00:14:34.673 }, 00:14:34.673 { 00:14:34.673 "name": "pt4", 00:14:34.673 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:34.673 "is_configured": true, 00:14:34.673 "data_offset": 2048, 00:14:34.673 "data_size": 63488 00:14:34.673 } 00:14:34.673 ] 00:14:34.673 }' 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.673 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.933 [2024-09-30 14:13:39.522250] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:34.933 [2024-09-30 14:13:39.522276] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:34.933 [2024-09-30 14:13:39.522339] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:34.933 [2024-09-30 14:13:39.522401] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:34.933 [2024-09-30 14:13:39.522412] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.933 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.193 [2024-09-30 14:13:39.594163] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:35.193 [2024-09-30 14:13:39.594218] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.193 [2024-09-30 14:13:39.594251] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:14:35.193 [2024-09-30 14:13:39.594261] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.193 [2024-09-30 14:13:39.596354] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.193 [2024-09-30 14:13:39.596394] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:35.193 [2024-09-30 14:13:39.596456] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:35.193 [2024-09-30 14:13:39.596512] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:35.193 [2024-09-30 14:13:39.596638] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:35.193 [2024-09-30 14:13:39.596657] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:35.193 [2024-09-30 14:13:39.596671] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:14:35.193 [2024-09-30 14:13:39.596713] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:35.193 [2024-09-30 14:13:39.596803] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:35.193 pt1 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.193 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.193 "name": "raid_bdev1", 00:14:35.193 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:35.193 "strip_size_kb": 64, 00:14:35.193 "state": "configuring", 00:14:35.193 "raid_level": "raid5f", 00:14:35.193 "superblock": true, 00:14:35.193 "num_base_bdevs": 4, 00:14:35.193 "num_base_bdevs_discovered": 2, 00:14:35.193 "num_base_bdevs_operational": 3, 00:14:35.193 "base_bdevs_list": [ 00:14:35.193 { 00:14:35.193 "name": null, 00:14:35.193 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.193 "is_configured": false, 00:14:35.193 "data_offset": 2048, 00:14:35.193 "data_size": 63488 00:14:35.193 }, 00:14:35.193 { 00:14:35.194 "name": "pt2", 00:14:35.194 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:35.194 "is_configured": true, 00:14:35.194 "data_offset": 2048, 00:14:35.194 "data_size": 63488 00:14:35.194 }, 00:14:35.194 { 00:14:35.194 "name": "pt3", 00:14:35.194 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:35.194 "is_configured": true, 00:14:35.194 "data_offset": 2048, 00:14:35.194 "data_size": 63488 00:14:35.194 }, 00:14:35.194 { 00:14:35.194 "name": null, 00:14:35.194 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:35.194 "is_configured": false, 00:14:35.194 "data_offset": 2048, 00:14:35.194 "data_size": 63488 00:14:35.194 } 00:14:35.194 ] 00:14:35.194 }' 00:14:35.194 14:13:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.194 14:13:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.453 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.453 [2024-09-30 14:13:40.069353] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:35.453 [2024-09-30 14:13:40.069447] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.453 [2024-09-30 14:13:40.069485] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:14:35.453 [2024-09-30 14:13:40.069493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.453 [2024-09-30 14:13:40.069813] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.453 [2024-09-30 14:13:40.069830] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:35.453 [2024-09-30 14:13:40.069883] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:35.453 [2024-09-30 14:13:40.069900] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:35.453 [2024-09-30 14:13:40.069984] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:14:35.454 [2024-09-30 14:13:40.069992] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:35.454 [2024-09-30 14:13:40.070208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:35.454 [2024-09-30 14:13:40.070704] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:14:35.454 [2024-09-30 14:13:40.070724] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:14:35.454 [2024-09-30 14:13:40.070878] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.454 pt4 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.454 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.713 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.713 "name": "raid_bdev1", 00:14:35.713 "uuid": "eb97dcad-06cd-4b5a-a6d6-5e669ea365e1", 00:14:35.713 "strip_size_kb": 64, 00:14:35.713 "state": "online", 00:14:35.713 "raid_level": "raid5f", 00:14:35.713 "superblock": true, 00:14:35.713 "num_base_bdevs": 4, 00:14:35.713 "num_base_bdevs_discovered": 3, 00:14:35.713 "num_base_bdevs_operational": 3, 00:14:35.713 "base_bdevs_list": [ 00:14:35.713 { 00:14:35.713 "name": null, 00:14:35.713 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.713 "is_configured": false, 00:14:35.713 "data_offset": 2048, 00:14:35.713 "data_size": 63488 00:14:35.713 }, 00:14:35.713 { 00:14:35.713 "name": "pt2", 00:14:35.713 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:35.713 "is_configured": true, 00:14:35.713 "data_offset": 2048, 00:14:35.713 "data_size": 63488 00:14:35.713 }, 00:14:35.713 { 00:14:35.713 "name": "pt3", 00:14:35.713 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:35.713 "is_configured": true, 00:14:35.713 "data_offset": 2048, 00:14:35.713 "data_size": 63488 00:14:35.713 }, 00:14:35.713 { 00:14:35.713 "name": "pt4", 00:14:35.713 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:35.713 "is_configured": true, 00:14:35.713 "data_offset": 2048, 00:14:35.713 "data_size": 63488 00:14:35.713 } 00:14:35.713 ] 00:14:35.713 }' 00:14:35.713 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.713 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.973 [2024-09-30 14:13:40.588668] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 '!=' eb97dcad-06cd-4b5a-a6d6-5e669ea365e1 ']' 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 95567 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 95567 ']' 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 95567 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:35.973 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 95567 00:14:36.233 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:36.233 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:36.233 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 95567' 00:14:36.233 killing process with pid 95567 00:14:36.233 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 95567 00:14:36.233 [2024-09-30 14:13:40.652666] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:36.233 [2024-09-30 14:13:40.652800] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:36.233 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 95567 00:14:36.233 [2024-09-30 14:13:40.652892] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:36.233 [2024-09-30 14:13:40.652914] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:14:36.233 [2024-09-30 14:13:40.695429] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:36.494 14:13:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:36.494 00:14:36.494 real 0m7.131s 00:14:36.494 user 0m11.952s 00:14:36.494 sys 0m1.583s 00:14:36.494 ************************************ 00:14:36.494 END TEST raid5f_superblock_test 00:14:36.494 ************************************ 00:14:36.494 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:36.494 14:13:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.494 14:13:41 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:36.494 14:13:41 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:14:36.494 14:13:41 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:36.494 14:13:41 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:36.494 14:13:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:36.494 ************************************ 00:14:36.494 START TEST raid5f_rebuild_test 00:14:36.494 ************************************ 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 false false true 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:36.494 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=96036 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 96036 00:14:36.495 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 96036 ']' 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:36.495 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.495 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:36.495 Zero copy mechanism will not be used. 00:14:36.495 [2024-09-30 14:13:41.130818] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:14:36.495 [2024-09-30 14:13:41.130961] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96036 ] 00:14:36.755 [2024-09-30 14:13:41.263827] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:14:36.755 [2024-09-30 14:13:41.293600] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:36.755 [2024-09-30 14:13:41.337694] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:36.755 [2024-09-30 14:13:41.379530] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:36.755 [2024-09-30 14:13:41.379573] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.325 BaseBdev1_malloc 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.325 [2024-09-30 14:13:41.961371] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:37.325 [2024-09-30 14:13:41.961468] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.325 [2024-09-30 14:13:41.961521] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:37.325 [2024-09-30 14:13:41.961540] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.325 [2024-09-30 14:13:41.963545] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.325 [2024-09-30 14:13:41.963579] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:37.325 BaseBdev1 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.325 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.585 BaseBdev2_malloc 00:14:37.586 14:13:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 [2024-09-30 14:13:42.007385] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:37.586 [2024-09-30 14:13:42.007577] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.586 [2024-09-30 14:13:42.007635] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:37.586 [2024-09-30 14:13:42.007671] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.586 [2024-09-30 14:13:42.012355] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.586 [2024-09-30 14:13:42.012430] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:37.586 BaseBdev2 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 BaseBdev3_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 [2024-09-30 14:13:42.038668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:37.586 [2024-09-30 14:13:42.038729] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.586 [2024-09-30 14:13:42.038749] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:37.586 [2024-09-30 14:13:42.038760] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.586 [2024-09-30 14:13:42.040840] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.586 [2024-09-30 14:13:42.040878] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:37.586 BaseBdev3 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 BaseBdev4_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 [2024-09-30 14:13:42.067214] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:37.586 [2024-09-30 14:13:42.067266] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.586 [2024-09-30 14:13:42.067284] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:37.586 [2024-09-30 14:13:42.067293] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.586 [2024-09-30 14:13:42.069312] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.586 [2024-09-30 14:13:42.069348] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:37.586 BaseBdev4 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 spare_malloc 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 spare_delay 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 [2024-09-30 14:13:42.107530] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:37.586 [2024-09-30 14:13:42.107584] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:37.586 [2024-09-30 14:13:42.107618] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:37.586 [2024-09-30 14:13:42.107629] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:37.586 [2024-09-30 14:13:42.109554] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:37.586 [2024-09-30 14:13:42.109589] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:37.586 spare 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 [2024-09-30 14:13:42.119583] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:37.586 [2024-09-30 14:13:42.121279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:37.586 [2024-09-30 14:13:42.121338] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:37.586 [2024-09-30 14:13:42.121375] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:37.586 [2024-09-30 14:13:42.121445] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:14:37.586 [2024-09-30 14:13:42.121457] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:37.586 [2024-09-30 14:13:42.121698] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:14:37.586 [2024-09-30 14:13:42.122087] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:14:37.586 [2024-09-30 14:13:42.122108] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:14:37.586 [2024-09-30 14:13:42.122218] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.586 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.586 "name": "raid_bdev1", 00:14:37.586 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:37.586 "strip_size_kb": 64, 00:14:37.586 "state": "online", 00:14:37.586 "raid_level": "raid5f", 00:14:37.586 "superblock": false, 00:14:37.586 "num_base_bdevs": 4, 00:14:37.586 "num_base_bdevs_discovered": 4, 00:14:37.586 "num_base_bdevs_operational": 4, 00:14:37.586 "base_bdevs_list": [ 00:14:37.586 { 00:14:37.586 "name": "BaseBdev1", 00:14:37.586 "uuid": "33ce3c4a-7699-5d5e-bbd9-ca1e0bb7045d", 00:14:37.586 "is_configured": true, 00:14:37.586 "data_offset": 0, 00:14:37.586 "data_size": 65536 00:14:37.586 }, 00:14:37.586 { 00:14:37.586 "name": "BaseBdev2", 00:14:37.586 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:37.586 "is_configured": true, 00:14:37.586 "data_offset": 0, 00:14:37.586 "data_size": 65536 00:14:37.586 }, 00:14:37.586 { 00:14:37.586 "name": "BaseBdev3", 00:14:37.586 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:37.586 "is_configured": true, 00:14:37.586 "data_offset": 0, 00:14:37.586 "data_size": 65536 00:14:37.586 }, 00:14:37.586 { 00:14:37.586 "name": "BaseBdev4", 00:14:37.586 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:37.586 "is_configured": true, 00:14:37.586 "data_offset": 0, 00:14:37.586 "data_size": 65536 00:14:37.586 } 00:14:37.586 ] 00:14:37.586 }' 00:14:37.587 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.587 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:38.157 [2024-09-30 14:13:42.587341] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:38.157 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:38.417 [2024-09-30 14:13:42.866728] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:38.417 /dev/nbd0 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:38.417 1+0 records in 00:14:38.417 1+0 records out 00:14:38.417 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000334637 s, 12.2 MB/s 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:38.417 14:13:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:14:38.987 512+0 records in 00:14:38.987 512+0 records out 00:14:38.987 100663296 bytes (101 MB, 96 MiB) copied, 0.46213 s, 218 MB/s 00:14:38.987 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:38.987 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:38.987 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:38.987 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:38.988 [2024-09-30 14:13:43.590649] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.988 [2024-09-30 14:13:43.626677] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.988 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.247 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.247 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.247 "name": "raid_bdev1", 00:14:39.247 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:39.247 "strip_size_kb": 64, 00:14:39.247 "state": "online", 00:14:39.247 "raid_level": "raid5f", 00:14:39.247 "superblock": false, 00:14:39.247 "num_base_bdevs": 4, 00:14:39.247 "num_base_bdevs_discovered": 3, 00:14:39.247 "num_base_bdevs_operational": 3, 00:14:39.247 "base_bdevs_list": [ 00:14:39.247 { 00:14:39.247 "name": null, 00:14:39.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.247 "is_configured": false, 00:14:39.247 "data_offset": 0, 00:14:39.247 "data_size": 65536 00:14:39.247 }, 00:14:39.247 { 00:14:39.247 "name": "BaseBdev2", 00:14:39.247 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 0, 00:14:39.247 "data_size": 65536 00:14:39.247 }, 00:14:39.247 { 00:14:39.247 "name": "BaseBdev3", 00:14:39.247 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 0, 00:14:39.247 "data_size": 65536 00:14:39.247 }, 00:14:39.247 { 00:14:39.247 "name": "BaseBdev4", 00:14:39.247 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 0, 00:14:39.247 "data_size": 65536 00:14:39.247 } 00:14:39.247 ] 00:14:39.247 }' 00:14:39.247 14:13:43 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.247 14:13:43 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.507 14:13:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:39.507 14:13:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.507 14:13:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.507 [2024-09-30 14:13:44.061883] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:39.507 [2024-09-30 14:13:44.065210] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b820 00:14:39.507 [2024-09-30 14:13:44.067235] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:39.507 14:13:44 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.507 14:13:44 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:40.445 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.446 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.704 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.704 "name": "raid_bdev1", 00:14:40.704 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:40.704 "strip_size_kb": 64, 00:14:40.704 "state": "online", 00:14:40.704 "raid_level": "raid5f", 00:14:40.704 "superblock": false, 00:14:40.704 "num_base_bdevs": 4, 00:14:40.704 "num_base_bdevs_discovered": 4, 00:14:40.704 "num_base_bdevs_operational": 4, 00:14:40.704 "process": { 00:14:40.704 "type": "rebuild", 00:14:40.704 "target": "spare", 00:14:40.704 "progress": { 00:14:40.704 "blocks": 19200, 00:14:40.704 "percent": 9 00:14:40.704 } 00:14:40.705 }, 00:14:40.705 "base_bdevs_list": [ 00:14:40.705 { 00:14:40.705 "name": "spare", 00:14:40.705 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev2", 00:14:40.705 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev3", 00:14:40.705 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev4", 00:14:40.705 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 } 00:14:40.705 ] 00:14:40.705 }' 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.705 [2024-09-30 14:13:45.229863] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:40.705 [2024-09-30 14:13:45.272501] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:40.705 [2024-09-30 14:13:45.272617] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:40.705 [2024-09-30 14:13:45.272635] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:40.705 [2024-09-30 14:13:45.272654] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.705 "name": "raid_bdev1", 00:14:40.705 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:40.705 "strip_size_kb": 64, 00:14:40.705 "state": "online", 00:14:40.705 "raid_level": "raid5f", 00:14:40.705 "superblock": false, 00:14:40.705 "num_base_bdevs": 4, 00:14:40.705 "num_base_bdevs_discovered": 3, 00:14:40.705 "num_base_bdevs_operational": 3, 00:14:40.705 "base_bdevs_list": [ 00:14:40.705 { 00:14:40.705 "name": null, 00:14:40.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.705 "is_configured": false, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev2", 00:14:40.705 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev3", 00:14:40.705 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 }, 00:14:40.705 { 00:14:40.705 "name": "BaseBdev4", 00:14:40.705 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:40.705 "is_configured": true, 00:14:40.705 "data_offset": 0, 00:14:40.705 "data_size": 65536 00:14:40.705 } 00:14:40.705 ] 00:14:40.705 }' 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.705 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.274 "name": "raid_bdev1", 00:14:41.274 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:41.274 "strip_size_kb": 64, 00:14:41.274 "state": "online", 00:14:41.274 "raid_level": "raid5f", 00:14:41.274 "superblock": false, 00:14:41.274 "num_base_bdevs": 4, 00:14:41.274 "num_base_bdevs_discovered": 3, 00:14:41.274 "num_base_bdevs_operational": 3, 00:14:41.274 "base_bdevs_list": [ 00:14:41.274 { 00:14:41.274 "name": null, 00:14:41.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.274 "is_configured": false, 00:14:41.274 "data_offset": 0, 00:14:41.274 "data_size": 65536 00:14:41.274 }, 00:14:41.274 { 00:14:41.274 "name": "BaseBdev2", 00:14:41.274 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:41.274 "is_configured": true, 00:14:41.274 "data_offset": 0, 00:14:41.274 "data_size": 65536 00:14:41.274 }, 00:14:41.274 { 00:14:41.274 "name": "BaseBdev3", 00:14:41.274 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:41.274 "is_configured": true, 00:14:41.274 "data_offset": 0, 00:14:41.274 "data_size": 65536 00:14:41.274 }, 00:14:41.274 { 00:14:41.274 "name": "BaseBdev4", 00:14:41.274 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:41.274 "is_configured": true, 00:14:41.274 "data_offset": 0, 00:14:41.274 "data_size": 65536 00:14:41.274 } 00:14:41.274 ] 00:14:41.274 }' 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.274 [2024-09-30 14:13:45.792849] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:41.274 [2024-09-30 14:13:45.796016] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b8f0 00:14:41.274 [2024-09-30 14:13:45.798039] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.274 14:13:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.210 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:42.210 "name": "raid_bdev1", 00:14:42.210 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:42.210 "strip_size_kb": 64, 00:14:42.210 "state": "online", 00:14:42.210 "raid_level": "raid5f", 00:14:42.210 "superblock": false, 00:14:42.210 "num_base_bdevs": 4, 00:14:42.210 "num_base_bdevs_discovered": 4, 00:14:42.210 "num_base_bdevs_operational": 4, 00:14:42.210 "process": { 00:14:42.210 "type": "rebuild", 00:14:42.210 "target": "spare", 00:14:42.210 "progress": { 00:14:42.210 "blocks": 19200, 00:14:42.210 "percent": 9 00:14:42.211 } 00:14:42.211 }, 00:14:42.211 "base_bdevs_list": [ 00:14:42.211 { 00:14:42.211 "name": "spare", 00:14:42.211 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:42.211 "is_configured": true, 00:14:42.211 "data_offset": 0, 00:14:42.211 "data_size": 65536 00:14:42.211 }, 00:14:42.211 { 00:14:42.211 "name": "BaseBdev2", 00:14:42.211 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:42.211 "is_configured": true, 00:14:42.211 "data_offset": 0, 00:14:42.211 "data_size": 65536 00:14:42.211 }, 00:14:42.211 { 00:14:42.211 "name": "BaseBdev3", 00:14:42.211 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:42.211 "is_configured": true, 00:14:42.211 "data_offset": 0, 00:14:42.211 "data_size": 65536 00:14:42.211 }, 00:14:42.211 { 00:14:42.211 "name": "BaseBdev4", 00:14:42.211 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:42.211 "is_configured": true, 00:14:42.211 "data_offset": 0, 00:14:42.211 "data_size": 65536 00:14:42.211 } 00:14:42.211 ] 00:14:42.211 }' 00:14:42.211 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=507 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 14:13:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:42.470 "name": "raid_bdev1", 00:14:42.470 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:42.470 "strip_size_kb": 64, 00:14:42.470 "state": "online", 00:14:42.470 "raid_level": "raid5f", 00:14:42.470 "superblock": false, 00:14:42.470 "num_base_bdevs": 4, 00:14:42.470 "num_base_bdevs_discovered": 4, 00:14:42.470 "num_base_bdevs_operational": 4, 00:14:42.470 "process": { 00:14:42.470 "type": "rebuild", 00:14:42.470 "target": "spare", 00:14:42.470 "progress": { 00:14:42.470 "blocks": 21120, 00:14:42.470 "percent": 10 00:14:42.470 } 00:14:42.470 }, 00:14:42.470 "base_bdevs_list": [ 00:14:42.470 { 00:14:42.470 "name": "spare", 00:14:42.470 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:42.470 "is_configured": true, 00:14:42.470 "data_offset": 0, 00:14:42.470 "data_size": 65536 00:14:42.470 }, 00:14:42.470 { 00:14:42.470 "name": "BaseBdev2", 00:14:42.470 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:42.470 "is_configured": true, 00:14:42.470 "data_offset": 0, 00:14:42.470 "data_size": 65536 00:14:42.470 }, 00:14:42.470 { 00:14:42.470 "name": "BaseBdev3", 00:14:42.470 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:42.470 "is_configured": true, 00:14:42.470 "data_offset": 0, 00:14:42.470 "data_size": 65536 00:14:42.470 }, 00:14:42.470 { 00:14:42.470 "name": "BaseBdev4", 00:14:42.470 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:42.470 "is_configured": true, 00:14:42.470 "data_offset": 0, 00:14:42.470 "data_size": 65536 00:14:42.470 } 00:14:42.470 ] 00:14:42.470 }' 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:42.470 14:13:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.846 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:43.847 "name": "raid_bdev1", 00:14:43.847 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:43.847 "strip_size_kb": 64, 00:14:43.847 "state": "online", 00:14:43.847 "raid_level": "raid5f", 00:14:43.847 "superblock": false, 00:14:43.847 "num_base_bdevs": 4, 00:14:43.847 "num_base_bdevs_discovered": 4, 00:14:43.847 "num_base_bdevs_operational": 4, 00:14:43.847 "process": { 00:14:43.847 "type": "rebuild", 00:14:43.847 "target": "spare", 00:14:43.847 "progress": { 00:14:43.847 "blocks": 42240, 00:14:43.847 "percent": 21 00:14:43.847 } 00:14:43.847 }, 00:14:43.847 "base_bdevs_list": [ 00:14:43.847 { 00:14:43.847 "name": "spare", 00:14:43.847 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:43.847 "is_configured": true, 00:14:43.847 "data_offset": 0, 00:14:43.847 "data_size": 65536 00:14:43.847 }, 00:14:43.847 { 00:14:43.847 "name": "BaseBdev2", 00:14:43.847 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:43.847 "is_configured": true, 00:14:43.847 "data_offset": 0, 00:14:43.847 "data_size": 65536 00:14:43.847 }, 00:14:43.847 { 00:14:43.847 "name": "BaseBdev3", 00:14:43.847 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:43.847 "is_configured": true, 00:14:43.847 "data_offset": 0, 00:14:43.847 "data_size": 65536 00:14:43.847 }, 00:14:43.847 { 00:14:43.847 "name": "BaseBdev4", 00:14:43.847 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:43.847 "is_configured": true, 00:14:43.847 "data_offset": 0, 00:14:43.847 "data_size": 65536 00:14:43.847 } 00:14:43.847 ] 00:14:43.847 }' 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:43.847 14:13:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:44.784 "name": "raid_bdev1", 00:14:44.784 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:44.784 "strip_size_kb": 64, 00:14:44.784 "state": "online", 00:14:44.784 "raid_level": "raid5f", 00:14:44.784 "superblock": false, 00:14:44.784 "num_base_bdevs": 4, 00:14:44.784 "num_base_bdevs_discovered": 4, 00:14:44.784 "num_base_bdevs_operational": 4, 00:14:44.784 "process": { 00:14:44.784 "type": "rebuild", 00:14:44.784 "target": "spare", 00:14:44.784 "progress": { 00:14:44.784 "blocks": 65280, 00:14:44.784 "percent": 33 00:14:44.784 } 00:14:44.784 }, 00:14:44.784 "base_bdevs_list": [ 00:14:44.784 { 00:14:44.784 "name": "spare", 00:14:44.784 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:44.784 "is_configured": true, 00:14:44.784 "data_offset": 0, 00:14:44.784 "data_size": 65536 00:14:44.784 }, 00:14:44.784 { 00:14:44.784 "name": "BaseBdev2", 00:14:44.784 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:44.784 "is_configured": true, 00:14:44.784 "data_offset": 0, 00:14:44.784 "data_size": 65536 00:14:44.784 }, 00:14:44.784 { 00:14:44.784 "name": "BaseBdev3", 00:14:44.784 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:44.784 "is_configured": true, 00:14:44.784 "data_offset": 0, 00:14:44.784 "data_size": 65536 00:14:44.784 }, 00:14:44.784 { 00:14:44.784 "name": "BaseBdev4", 00:14:44.784 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:44.784 "is_configured": true, 00:14:44.784 "data_offset": 0, 00:14:44.784 "data_size": 65536 00:14:44.784 } 00:14:44.784 ] 00:14:44.784 }' 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:44.784 14:13:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.721 14:13:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.980 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:45.980 "name": "raid_bdev1", 00:14:45.980 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:45.980 "strip_size_kb": 64, 00:14:45.980 "state": "online", 00:14:45.980 "raid_level": "raid5f", 00:14:45.980 "superblock": false, 00:14:45.980 "num_base_bdevs": 4, 00:14:45.980 "num_base_bdevs_discovered": 4, 00:14:45.980 "num_base_bdevs_operational": 4, 00:14:45.980 "process": { 00:14:45.980 "type": "rebuild", 00:14:45.980 "target": "spare", 00:14:45.980 "progress": { 00:14:45.980 "blocks": 86400, 00:14:45.980 "percent": 43 00:14:45.980 } 00:14:45.980 }, 00:14:45.980 "base_bdevs_list": [ 00:14:45.980 { 00:14:45.980 "name": "spare", 00:14:45.980 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:45.980 "is_configured": true, 00:14:45.980 "data_offset": 0, 00:14:45.980 "data_size": 65536 00:14:45.980 }, 00:14:45.980 { 00:14:45.980 "name": "BaseBdev2", 00:14:45.980 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:45.980 "is_configured": true, 00:14:45.980 "data_offset": 0, 00:14:45.980 "data_size": 65536 00:14:45.980 }, 00:14:45.980 { 00:14:45.980 "name": "BaseBdev3", 00:14:45.980 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:45.980 "is_configured": true, 00:14:45.980 "data_offset": 0, 00:14:45.980 "data_size": 65536 00:14:45.980 }, 00:14:45.980 { 00:14:45.980 "name": "BaseBdev4", 00:14:45.980 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:45.980 "is_configured": true, 00:14:45.980 "data_offset": 0, 00:14:45.980 "data_size": 65536 00:14:45.980 } 00:14:45.980 ] 00:14:45.980 }' 00:14:45.980 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:45.980 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:45.981 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:45.981 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:45.981 14:13:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:46.918 "name": "raid_bdev1", 00:14:46.918 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:46.918 "strip_size_kb": 64, 00:14:46.918 "state": "online", 00:14:46.918 "raid_level": "raid5f", 00:14:46.918 "superblock": false, 00:14:46.918 "num_base_bdevs": 4, 00:14:46.918 "num_base_bdevs_discovered": 4, 00:14:46.918 "num_base_bdevs_operational": 4, 00:14:46.918 "process": { 00:14:46.918 "type": "rebuild", 00:14:46.918 "target": "spare", 00:14:46.918 "progress": { 00:14:46.918 "blocks": 107520, 00:14:46.918 "percent": 54 00:14:46.918 } 00:14:46.918 }, 00:14:46.918 "base_bdevs_list": [ 00:14:46.918 { 00:14:46.918 "name": "spare", 00:14:46.918 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:46.918 "is_configured": true, 00:14:46.918 "data_offset": 0, 00:14:46.918 "data_size": 65536 00:14:46.918 }, 00:14:46.918 { 00:14:46.918 "name": "BaseBdev2", 00:14:46.918 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:46.918 "is_configured": true, 00:14:46.918 "data_offset": 0, 00:14:46.918 "data_size": 65536 00:14:46.918 }, 00:14:46.918 { 00:14:46.918 "name": "BaseBdev3", 00:14:46.918 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:46.918 "is_configured": true, 00:14:46.918 "data_offset": 0, 00:14:46.918 "data_size": 65536 00:14:46.918 }, 00:14:46.918 { 00:14:46.918 "name": "BaseBdev4", 00:14:46.918 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:46.918 "is_configured": true, 00:14:46.918 "data_offset": 0, 00:14:46.918 "data_size": 65536 00:14:46.918 } 00:14:46.918 ] 00:14:46.918 }' 00:14:46.918 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:46.919 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:46.919 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:47.178 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:47.178 14:13:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:48.115 "name": "raid_bdev1", 00:14:48.115 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:48.115 "strip_size_kb": 64, 00:14:48.115 "state": "online", 00:14:48.115 "raid_level": "raid5f", 00:14:48.115 "superblock": false, 00:14:48.115 "num_base_bdevs": 4, 00:14:48.115 "num_base_bdevs_discovered": 4, 00:14:48.115 "num_base_bdevs_operational": 4, 00:14:48.115 "process": { 00:14:48.115 "type": "rebuild", 00:14:48.115 "target": "spare", 00:14:48.115 "progress": { 00:14:48.115 "blocks": 130560, 00:14:48.115 "percent": 66 00:14:48.115 } 00:14:48.115 }, 00:14:48.115 "base_bdevs_list": [ 00:14:48.115 { 00:14:48.115 "name": "spare", 00:14:48.115 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:48.115 "is_configured": true, 00:14:48.115 "data_offset": 0, 00:14:48.115 "data_size": 65536 00:14:48.115 }, 00:14:48.115 { 00:14:48.115 "name": "BaseBdev2", 00:14:48.115 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:48.115 "is_configured": true, 00:14:48.115 "data_offset": 0, 00:14:48.115 "data_size": 65536 00:14:48.115 }, 00:14:48.115 { 00:14:48.115 "name": "BaseBdev3", 00:14:48.115 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:48.115 "is_configured": true, 00:14:48.115 "data_offset": 0, 00:14:48.115 "data_size": 65536 00:14:48.115 }, 00:14:48.115 { 00:14:48.115 "name": "BaseBdev4", 00:14:48.115 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:48.115 "is_configured": true, 00:14:48.115 "data_offset": 0, 00:14:48.115 "data_size": 65536 00:14:48.115 } 00:14:48.115 ] 00:14:48.115 }' 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:48.115 14:13:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:49.496 "name": "raid_bdev1", 00:14:49.496 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:49.496 "strip_size_kb": 64, 00:14:49.496 "state": "online", 00:14:49.496 "raid_level": "raid5f", 00:14:49.496 "superblock": false, 00:14:49.496 "num_base_bdevs": 4, 00:14:49.496 "num_base_bdevs_discovered": 4, 00:14:49.496 "num_base_bdevs_operational": 4, 00:14:49.496 "process": { 00:14:49.496 "type": "rebuild", 00:14:49.496 "target": "spare", 00:14:49.496 "progress": { 00:14:49.496 "blocks": 151680, 00:14:49.496 "percent": 77 00:14:49.496 } 00:14:49.496 }, 00:14:49.496 "base_bdevs_list": [ 00:14:49.496 { 00:14:49.496 "name": "spare", 00:14:49.496 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:49.496 "is_configured": true, 00:14:49.496 "data_offset": 0, 00:14:49.496 "data_size": 65536 00:14:49.496 }, 00:14:49.496 { 00:14:49.496 "name": "BaseBdev2", 00:14:49.496 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:49.496 "is_configured": true, 00:14:49.496 "data_offset": 0, 00:14:49.496 "data_size": 65536 00:14:49.496 }, 00:14:49.496 { 00:14:49.496 "name": "BaseBdev3", 00:14:49.496 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:49.496 "is_configured": true, 00:14:49.496 "data_offset": 0, 00:14:49.496 "data_size": 65536 00:14:49.496 }, 00:14:49.496 { 00:14:49.496 "name": "BaseBdev4", 00:14:49.496 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:49.496 "is_configured": true, 00:14:49.496 "data_offset": 0, 00:14:49.496 "data_size": 65536 00:14:49.496 } 00:14:49.496 ] 00:14:49.496 }' 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:49.496 14:13:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:50.435 "name": "raid_bdev1", 00:14:50.435 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:50.435 "strip_size_kb": 64, 00:14:50.435 "state": "online", 00:14:50.435 "raid_level": "raid5f", 00:14:50.435 "superblock": false, 00:14:50.435 "num_base_bdevs": 4, 00:14:50.435 "num_base_bdevs_discovered": 4, 00:14:50.435 "num_base_bdevs_operational": 4, 00:14:50.435 "process": { 00:14:50.435 "type": "rebuild", 00:14:50.435 "target": "spare", 00:14:50.435 "progress": { 00:14:50.435 "blocks": 174720, 00:14:50.435 "percent": 88 00:14:50.435 } 00:14:50.435 }, 00:14:50.435 "base_bdevs_list": [ 00:14:50.435 { 00:14:50.435 "name": "spare", 00:14:50.435 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:50.435 "is_configured": true, 00:14:50.435 "data_offset": 0, 00:14:50.435 "data_size": 65536 00:14:50.435 }, 00:14:50.435 { 00:14:50.435 "name": "BaseBdev2", 00:14:50.435 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:50.435 "is_configured": true, 00:14:50.435 "data_offset": 0, 00:14:50.435 "data_size": 65536 00:14:50.435 }, 00:14:50.435 { 00:14:50.435 "name": "BaseBdev3", 00:14:50.435 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:50.435 "is_configured": true, 00:14:50.435 "data_offset": 0, 00:14:50.435 "data_size": 65536 00:14:50.435 }, 00:14:50.435 { 00:14:50.435 "name": "BaseBdev4", 00:14:50.435 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:50.435 "is_configured": true, 00:14:50.435 "data_offset": 0, 00:14:50.435 "data_size": 65536 00:14:50.435 } 00:14:50.435 ] 00:14:50.435 }' 00:14:50.435 14:13:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:50.435 14:13:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:50.435 14:13:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:50.435 14:13:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:50.435 14:13:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:51.890 "name": "raid_bdev1", 00:14:51.890 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:51.890 "strip_size_kb": 64, 00:14:51.890 "state": "online", 00:14:51.890 "raid_level": "raid5f", 00:14:51.890 "superblock": false, 00:14:51.890 "num_base_bdevs": 4, 00:14:51.890 "num_base_bdevs_discovered": 4, 00:14:51.890 "num_base_bdevs_operational": 4, 00:14:51.890 "process": { 00:14:51.890 "type": "rebuild", 00:14:51.890 "target": "spare", 00:14:51.890 "progress": { 00:14:51.890 "blocks": 195840, 00:14:51.890 "percent": 99 00:14:51.890 } 00:14:51.890 }, 00:14:51.890 "base_bdevs_list": [ 00:14:51.890 { 00:14:51.890 "name": "spare", 00:14:51.890 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:51.890 "is_configured": true, 00:14:51.890 "data_offset": 0, 00:14:51.890 "data_size": 65536 00:14:51.890 }, 00:14:51.890 { 00:14:51.890 "name": "BaseBdev2", 00:14:51.890 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:51.890 "is_configured": true, 00:14:51.890 "data_offset": 0, 00:14:51.890 "data_size": 65536 00:14:51.890 }, 00:14:51.890 { 00:14:51.890 "name": "BaseBdev3", 00:14:51.890 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:51.890 "is_configured": true, 00:14:51.890 "data_offset": 0, 00:14:51.890 "data_size": 65536 00:14:51.890 }, 00:14:51.890 { 00:14:51.890 "name": "BaseBdev4", 00:14:51.890 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:51.890 "is_configured": true, 00:14:51.890 "data_offset": 0, 00:14:51.890 "data_size": 65536 00:14:51.890 } 00:14:51.890 ] 00:14:51.890 }' 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:51.890 [2024-09-30 14:13:56.137471] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:51.890 [2024-09-30 14:13:56.137596] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:51.890 [2024-09-30 14:13:56.137666] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:51.890 14:13:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:52.834 "name": "raid_bdev1", 00:14:52.834 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:52.834 "strip_size_kb": 64, 00:14:52.834 "state": "online", 00:14:52.834 "raid_level": "raid5f", 00:14:52.834 "superblock": false, 00:14:52.834 "num_base_bdevs": 4, 00:14:52.834 "num_base_bdevs_discovered": 4, 00:14:52.834 "num_base_bdevs_operational": 4, 00:14:52.834 "base_bdevs_list": [ 00:14:52.834 { 00:14:52.834 "name": "spare", 00:14:52.834 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev2", 00:14:52.834 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev3", 00:14:52.834 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev4", 00:14:52.834 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 } 00:14:52.834 ] 00:14:52.834 }' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:52.834 "name": "raid_bdev1", 00:14:52.834 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:52.834 "strip_size_kb": 64, 00:14:52.834 "state": "online", 00:14:52.834 "raid_level": "raid5f", 00:14:52.834 "superblock": false, 00:14:52.834 "num_base_bdevs": 4, 00:14:52.834 "num_base_bdevs_discovered": 4, 00:14:52.834 "num_base_bdevs_operational": 4, 00:14:52.834 "base_bdevs_list": [ 00:14:52.834 { 00:14:52.834 "name": "spare", 00:14:52.834 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev2", 00:14:52.834 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev3", 00:14:52.834 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 }, 00:14:52.834 { 00:14:52.834 "name": "BaseBdev4", 00:14:52.834 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:52.834 "is_configured": true, 00:14:52.834 "data_offset": 0, 00:14:52.834 "data_size": 65536 00:14:52.834 } 00:14:52.834 ] 00:14:52.834 }' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:52.834 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.094 "name": "raid_bdev1", 00:14:53.094 "uuid": "44d5716e-b519-44fb-b24e-2ef0d2125955", 00:14:53.094 "strip_size_kb": 64, 00:14:53.094 "state": "online", 00:14:53.094 "raid_level": "raid5f", 00:14:53.094 "superblock": false, 00:14:53.094 "num_base_bdevs": 4, 00:14:53.094 "num_base_bdevs_discovered": 4, 00:14:53.094 "num_base_bdevs_operational": 4, 00:14:53.094 "base_bdevs_list": [ 00:14:53.094 { 00:14:53.094 "name": "spare", 00:14:53.094 "uuid": "fc19e4c3-2300-51ca-b9bf-b2df52c38e76", 00:14:53.094 "is_configured": true, 00:14:53.094 "data_offset": 0, 00:14:53.094 "data_size": 65536 00:14:53.094 }, 00:14:53.094 { 00:14:53.094 "name": "BaseBdev2", 00:14:53.094 "uuid": "b1787f86-475f-5c12-9d4f-dd5ff10a770b", 00:14:53.094 "is_configured": true, 00:14:53.094 "data_offset": 0, 00:14:53.094 "data_size": 65536 00:14:53.094 }, 00:14:53.094 { 00:14:53.094 "name": "BaseBdev3", 00:14:53.094 "uuid": "671ec74a-74f9-541e-a1b4-cf736aafdf72", 00:14:53.094 "is_configured": true, 00:14:53.094 "data_offset": 0, 00:14:53.094 "data_size": 65536 00:14:53.094 }, 00:14:53.094 { 00:14:53.094 "name": "BaseBdev4", 00:14:53.094 "uuid": "1b1d4add-851e-5677-b3dc-b91353985ab8", 00:14:53.094 "is_configured": true, 00:14:53.094 "data_offset": 0, 00:14:53.094 "data_size": 65536 00:14:53.094 } 00:14:53.094 ] 00:14:53.094 }' 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.094 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.354 [2024-09-30 14:13:57.991585] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:53.354 [2024-09-30 14:13:57.991665] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:53.354 [2024-09-30 14:13:57.991758] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:53.354 [2024-09-30 14:13:57.991849] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:53.354 [2024-09-30 14:13:57.991859] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.354 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.354 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.354 14:13:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:53.614 /dev/nbd0 00:14:53.614 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:53.875 1+0 records in 00:14:53.875 1+0 records out 00:14:53.875 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000303267 s, 13.5 MB/s 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:53.875 /dev/nbd1 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:53.875 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:54.135 1+0 records in 00:14:54.135 1+0 records out 00:14:54.135 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000423645 s, 9.7 MB/s 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:54.135 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:54.396 14:13:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 96036 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 96036 ']' 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 96036 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:54.396 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96036 00:14:54.655 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:54.655 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:54.655 killing process with pid 96036 00:14:54.655 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96036' 00:14:54.655 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 96036 00:14:54.655 Received shutdown signal, test time was about 60.000000 seconds 00:14:54.655 00:14:54.655 Latency(us) 00:14:54.655 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:54.655 =================================================================================================================== 00:14:54.655 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:54.655 [2024-09-30 14:13:59.065207] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:54.655 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 96036 00:14:54.655 [2024-09-30 14:13:59.114301] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:54.915 ************************************ 00:14:54.915 END TEST raid5f_rebuild_test 00:14:54.915 ************************************ 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:54.915 00:14:54.915 real 0m18.319s 00:14:54.915 user 0m22.076s 00:14:54.915 sys 0m2.280s 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.915 14:13:59 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:14:54.915 14:13:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:54.915 14:13:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:54.915 14:13:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:54.915 ************************************ 00:14:54.915 START TEST raid5f_rebuild_test_sb 00:14:54.915 ************************************ 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 true false true 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:54.915 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=96543 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 96543 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 96543 ']' 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:54.916 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:54.916 14:13:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.916 [2024-09-30 14:13:59.520474] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:14:54.916 [2024-09-30 14:13:59.520663] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:14:54.916 Zero copy mechanism will not be used. 00:14:54.916 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96543 ] 00:14:55.176 [2024-09-30 14:13:59.652370] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:14:55.176 [2024-09-30 14:13:59.681583] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:55.176 [2024-09-30 14:13:59.729086] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:55.176 [2024-09-30 14:13:59.771406] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:55.176 [2024-09-30 14:13:59.771542] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.746 BaseBdev1_malloc 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.746 [2024-09-30 14:14:00.353409] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:55.746 [2024-09-30 14:14:00.353594] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:55.746 [2024-09-30 14:14:00.353630] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:55.746 [2024-09-30 14:14:00.353651] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:55.746 [2024-09-30 14:14:00.355618] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:55.746 [2024-09-30 14:14:00.355654] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:55.746 BaseBdev1 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.746 BaseBdev2_malloc 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.746 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.746 [2024-09-30 14:14:00.399243] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:55.746 [2024-09-30 14:14:00.399368] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:55.746 [2024-09-30 14:14:00.399411] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:55.746 [2024-09-30 14:14:00.399436] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.007 [2024-09-30 14:14:00.404054] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.007 [2024-09-30 14:14:00.404109] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:56.007 BaseBdev2 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 BaseBdev3_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 [2024-09-30 14:14:00.430098] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:56.007 [2024-09-30 14:14:00.430234] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.007 [2024-09-30 14:14:00.430271] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:56.007 [2024-09-30 14:14:00.430300] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.007 [2024-09-30 14:14:00.432273] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.007 [2024-09-30 14:14:00.432352] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:56.007 BaseBdev3 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 BaseBdev4_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 [2024-09-30 14:14:00.458627] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:56.007 [2024-09-30 14:14:00.458722] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.007 [2024-09-30 14:14:00.458742] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:56.007 [2024-09-30 14:14:00.458753] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.007 [2024-09-30 14:14:00.460707] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.007 [2024-09-30 14:14:00.460742] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:56.007 BaseBdev4 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 spare_malloc 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 spare_delay 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 [2024-09-30 14:14:00.498888] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:56.007 [2024-09-30 14:14:00.498940] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.007 [2024-09-30 14:14:00.498958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:56.007 [2024-09-30 14:14:00.498969] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.007 [2024-09-30 14:14:00.500890] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.007 [2024-09-30 14:14:00.500980] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:56.007 spare 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 [2024-09-30 14:14:00.510949] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:56.007 [2024-09-30 14:14:00.512640] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:56.007 [2024-09-30 14:14:00.512699] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:56.007 [2024-09-30 14:14:00.512737] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:56.007 [2024-09-30 14:14:00.512885] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:14:56.007 [2024-09-30 14:14:00.512900] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:56.007 [2024-09-30 14:14:00.513124] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:14:56.007 [2024-09-30 14:14:00.513550] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:14:56.007 [2024-09-30 14:14:00.513561] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:14:56.007 [2024-09-30 14:14:00.513671] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.007 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.007 "name": "raid_bdev1", 00:14:56.007 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:14:56.007 "strip_size_kb": 64, 00:14:56.007 "state": "online", 00:14:56.007 "raid_level": "raid5f", 00:14:56.007 "superblock": true, 00:14:56.007 "num_base_bdevs": 4, 00:14:56.007 "num_base_bdevs_discovered": 4, 00:14:56.007 "num_base_bdevs_operational": 4, 00:14:56.007 "base_bdevs_list": [ 00:14:56.007 { 00:14:56.007 "name": "BaseBdev1", 00:14:56.007 "uuid": "8adc3a3d-afe7-5ec4-b73e-a54bef1d1c41", 00:14:56.007 "is_configured": true, 00:14:56.007 "data_offset": 2048, 00:14:56.007 "data_size": 63488 00:14:56.007 }, 00:14:56.007 { 00:14:56.007 "name": "BaseBdev2", 00:14:56.007 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:14:56.007 "is_configured": true, 00:14:56.007 "data_offset": 2048, 00:14:56.007 "data_size": 63488 00:14:56.007 }, 00:14:56.007 { 00:14:56.007 "name": "BaseBdev3", 00:14:56.007 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:14:56.007 "is_configured": true, 00:14:56.007 "data_offset": 2048, 00:14:56.007 "data_size": 63488 00:14:56.007 }, 00:14:56.007 { 00:14:56.007 "name": "BaseBdev4", 00:14:56.007 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:14:56.007 "is_configured": true, 00:14:56.007 "data_offset": 2048, 00:14:56.007 "data_size": 63488 00:14:56.008 } 00:14:56.008 ] 00:14:56.008 }' 00:14:56.008 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.008 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.267 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:56.267 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:56.267 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.267 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.267 [2024-09-30 14:14:00.910785] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.528 14:14:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:56.528 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:56.528 [2024-09-30 14:14:01.182253] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:56.788 /dev/nbd0 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:56.788 1+0 records in 00:14:56.788 1+0 records out 00:14:56.788 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000515222 s, 7.9 MB/s 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:56.788 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:14:57.048 496+0 records in 00:14:57.048 496+0 records out 00:14:57.048 97517568 bytes (98 MB, 93 MiB) copied, 0.380616 s, 256 MB/s 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:57.048 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:57.308 [2024-09-30 14:14:01.841442] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.308 [2024-09-30 14:14:01.857512] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.308 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.309 "name": "raid_bdev1", 00:14:57.309 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:14:57.309 "strip_size_kb": 64, 00:14:57.309 "state": "online", 00:14:57.309 "raid_level": "raid5f", 00:14:57.309 "superblock": true, 00:14:57.309 "num_base_bdevs": 4, 00:14:57.309 "num_base_bdevs_discovered": 3, 00:14:57.309 "num_base_bdevs_operational": 3, 00:14:57.309 "base_bdevs_list": [ 00:14:57.309 { 00:14:57.309 "name": null, 00:14:57.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.309 "is_configured": false, 00:14:57.309 "data_offset": 0, 00:14:57.309 "data_size": 63488 00:14:57.309 }, 00:14:57.309 { 00:14:57.309 "name": "BaseBdev2", 00:14:57.309 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:14:57.309 "is_configured": true, 00:14:57.309 "data_offset": 2048, 00:14:57.309 "data_size": 63488 00:14:57.309 }, 00:14:57.309 { 00:14:57.309 "name": "BaseBdev3", 00:14:57.309 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:14:57.309 "is_configured": true, 00:14:57.309 "data_offset": 2048, 00:14:57.309 "data_size": 63488 00:14:57.309 }, 00:14:57.309 { 00:14:57.309 "name": "BaseBdev4", 00:14:57.309 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:14:57.309 "is_configured": true, 00:14:57.309 "data_offset": 2048, 00:14:57.309 "data_size": 63488 00:14:57.309 } 00:14:57.309 ] 00:14:57.309 }' 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.309 14:14:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.878 14:14:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:57.879 14:14:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.879 14:14:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.879 [2024-09-30 14:14:02.336659] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:57.879 [2024-09-30 14:14:02.340063] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002ab20 00:14:57.879 [2024-09-30 14:14:02.342177] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:57.879 14:14:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.879 14:14:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.817 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.818 "name": "raid_bdev1", 00:14:58.818 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:14:58.818 "strip_size_kb": 64, 00:14:58.818 "state": "online", 00:14:58.818 "raid_level": "raid5f", 00:14:58.818 "superblock": true, 00:14:58.818 "num_base_bdevs": 4, 00:14:58.818 "num_base_bdevs_discovered": 4, 00:14:58.818 "num_base_bdevs_operational": 4, 00:14:58.818 "process": { 00:14:58.818 "type": "rebuild", 00:14:58.818 "target": "spare", 00:14:58.818 "progress": { 00:14:58.818 "blocks": 19200, 00:14:58.818 "percent": 10 00:14:58.818 } 00:14:58.818 }, 00:14:58.818 "base_bdevs_list": [ 00:14:58.818 { 00:14:58.818 "name": "spare", 00:14:58.818 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:14:58.818 "is_configured": true, 00:14:58.818 "data_offset": 2048, 00:14:58.818 "data_size": 63488 00:14:58.818 }, 00:14:58.818 { 00:14:58.818 "name": "BaseBdev2", 00:14:58.818 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:14:58.818 "is_configured": true, 00:14:58.818 "data_offset": 2048, 00:14:58.818 "data_size": 63488 00:14:58.818 }, 00:14:58.818 { 00:14:58.818 "name": "BaseBdev3", 00:14:58.818 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:14:58.818 "is_configured": true, 00:14:58.818 "data_offset": 2048, 00:14:58.818 "data_size": 63488 00:14:58.818 }, 00:14:58.818 { 00:14:58.818 "name": "BaseBdev4", 00:14:58.818 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:14:58.818 "is_configured": true, 00:14:58.818 "data_offset": 2048, 00:14:58.818 "data_size": 63488 00:14:58.818 } 00:14:58.818 ] 00:14:58.818 }' 00:14:58.818 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.818 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:58.818 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.077 [2024-09-30 14:14:03.513284] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:59.077 [2024-09-30 14:14:03.547381] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:59.077 [2024-09-30 14:14:03.547506] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:59.077 [2024-09-30 14:14:03.547524] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:59.077 [2024-09-30 14:14:03.547536] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:59.077 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:59.078 "name": "raid_bdev1", 00:14:59.078 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:14:59.078 "strip_size_kb": 64, 00:14:59.078 "state": "online", 00:14:59.078 "raid_level": "raid5f", 00:14:59.078 "superblock": true, 00:14:59.078 "num_base_bdevs": 4, 00:14:59.078 "num_base_bdevs_discovered": 3, 00:14:59.078 "num_base_bdevs_operational": 3, 00:14:59.078 "base_bdevs_list": [ 00:14:59.078 { 00:14:59.078 "name": null, 00:14:59.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:59.078 "is_configured": false, 00:14:59.078 "data_offset": 0, 00:14:59.078 "data_size": 63488 00:14:59.078 }, 00:14:59.078 { 00:14:59.078 "name": "BaseBdev2", 00:14:59.078 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:14:59.078 "is_configured": true, 00:14:59.078 "data_offset": 2048, 00:14:59.078 "data_size": 63488 00:14:59.078 }, 00:14:59.078 { 00:14:59.078 "name": "BaseBdev3", 00:14:59.078 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:14:59.078 "is_configured": true, 00:14:59.078 "data_offset": 2048, 00:14:59.078 "data_size": 63488 00:14:59.078 }, 00:14:59.078 { 00:14:59.078 "name": "BaseBdev4", 00:14:59.078 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:14:59.078 "is_configured": true, 00:14:59.078 "data_offset": 2048, 00:14:59.078 "data_size": 63488 00:14:59.078 } 00:14:59.078 ] 00:14:59.078 }' 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:59.078 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.337 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:59.337 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:59.337 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:59.337 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:59.597 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:59.597 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:59.597 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.597 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.597 14:14:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:59.597 "name": "raid_bdev1", 00:14:59.597 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:14:59.597 "strip_size_kb": 64, 00:14:59.597 "state": "online", 00:14:59.597 "raid_level": "raid5f", 00:14:59.597 "superblock": true, 00:14:59.597 "num_base_bdevs": 4, 00:14:59.597 "num_base_bdevs_discovered": 3, 00:14:59.597 "num_base_bdevs_operational": 3, 00:14:59.597 "base_bdevs_list": [ 00:14:59.597 { 00:14:59.597 "name": null, 00:14:59.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:59.597 "is_configured": false, 00:14:59.597 "data_offset": 0, 00:14:59.597 "data_size": 63488 00:14:59.597 }, 00:14:59.597 { 00:14:59.597 "name": "BaseBdev2", 00:14:59.597 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:14:59.597 "is_configured": true, 00:14:59.597 "data_offset": 2048, 00:14:59.597 "data_size": 63488 00:14:59.597 }, 00:14:59.597 { 00:14:59.597 "name": "BaseBdev3", 00:14:59.597 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:14:59.597 "is_configured": true, 00:14:59.597 "data_offset": 2048, 00:14:59.597 "data_size": 63488 00:14:59.597 }, 00:14:59.597 { 00:14:59.597 "name": "BaseBdev4", 00:14:59.597 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:14:59.597 "is_configured": true, 00:14:59.597 "data_offset": 2048, 00:14:59.597 "data_size": 63488 00:14:59.597 } 00:14:59.597 ] 00:14:59.597 }' 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.597 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.597 [2024-09-30 14:14:04.119765] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:59.597 [2024-09-30 14:14:04.122444] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002abf0 00:14:59.597 [2024-09-30 14:14:04.124521] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:59.598 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.598 14:14:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.538 "name": "raid_bdev1", 00:15:00.538 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:00.538 "strip_size_kb": 64, 00:15:00.538 "state": "online", 00:15:00.538 "raid_level": "raid5f", 00:15:00.538 "superblock": true, 00:15:00.538 "num_base_bdevs": 4, 00:15:00.538 "num_base_bdevs_discovered": 4, 00:15:00.538 "num_base_bdevs_operational": 4, 00:15:00.538 "process": { 00:15:00.538 "type": "rebuild", 00:15:00.538 "target": "spare", 00:15:00.538 "progress": { 00:15:00.538 "blocks": 19200, 00:15:00.538 "percent": 10 00:15:00.538 } 00:15:00.538 }, 00:15:00.538 "base_bdevs_list": [ 00:15:00.538 { 00:15:00.538 "name": "spare", 00:15:00.538 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:00.538 "is_configured": true, 00:15:00.538 "data_offset": 2048, 00:15:00.538 "data_size": 63488 00:15:00.538 }, 00:15:00.538 { 00:15:00.538 "name": "BaseBdev2", 00:15:00.538 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:00.538 "is_configured": true, 00:15:00.538 "data_offset": 2048, 00:15:00.538 "data_size": 63488 00:15:00.538 }, 00:15:00.538 { 00:15:00.538 "name": "BaseBdev3", 00:15:00.538 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:00.538 "is_configured": true, 00:15:00.538 "data_offset": 2048, 00:15:00.538 "data_size": 63488 00:15:00.538 }, 00:15:00.538 { 00:15:00.538 "name": "BaseBdev4", 00:15:00.538 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:00.538 "is_configured": true, 00:15:00.538 "data_offset": 2048, 00:15:00.538 "data_size": 63488 00:15:00.538 } 00:15:00.538 ] 00:15:00.538 }' 00:15:00.538 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:00.799 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=526 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.799 "name": "raid_bdev1", 00:15:00.799 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:00.799 "strip_size_kb": 64, 00:15:00.799 "state": "online", 00:15:00.799 "raid_level": "raid5f", 00:15:00.799 "superblock": true, 00:15:00.799 "num_base_bdevs": 4, 00:15:00.799 "num_base_bdevs_discovered": 4, 00:15:00.799 "num_base_bdevs_operational": 4, 00:15:00.799 "process": { 00:15:00.799 "type": "rebuild", 00:15:00.799 "target": "spare", 00:15:00.799 "progress": { 00:15:00.799 "blocks": 21120, 00:15:00.799 "percent": 11 00:15:00.799 } 00:15:00.799 }, 00:15:00.799 "base_bdevs_list": [ 00:15:00.799 { 00:15:00.799 "name": "spare", 00:15:00.799 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:00.799 "is_configured": true, 00:15:00.799 "data_offset": 2048, 00:15:00.799 "data_size": 63488 00:15:00.799 }, 00:15:00.799 { 00:15:00.799 "name": "BaseBdev2", 00:15:00.799 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:00.799 "is_configured": true, 00:15:00.799 "data_offset": 2048, 00:15:00.799 "data_size": 63488 00:15:00.799 }, 00:15:00.799 { 00:15:00.799 "name": "BaseBdev3", 00:15:00.799 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:00.799 "is_configured": true, 00:15:00.799 "data_offset": 2048, 00:15:00.799 "data_size": 63488 00:15:00.799 }, 00:15:00.799 { 00:15:00.799 "name": "BaseBdev4", 00:15:00.799 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:00.799 "is_configured": true, 00:15:00.799 "data_offset": 2048, 00:15:00.799 "data_size": 63488 00:15:00.799 } 00:15:00.799 ] 00:15:00.799 }' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.799 14:14:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:02.178 "name": "raid_bdev1", 00:15:02.178 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:02.178 "strip_size_kb": 64, 00:15:02.178 "state": "online", 00:15:02.178 "raid_level": "raid5f", 00:15:02.178 "superblock": true, 00:15:02.178 "num_base_bdevs": 4, 00:15:02.178 "num_base_bdevs_discovered": 4, 00:15:02.178 "num_base_bdevs_operational": 4, 00:15:02.178 "process": { 00:15:02.178 "type": "rebuild", 00:15:02.178 "target": "spare", 00:15:02.178 "progress": { 00:15:02.178 "blocks": 44160, 00:15:02.178 "percent": 23 00:15:02.178 } 00:15:02.178 }, 00:15:02.178 "base_bdevs_list": [ 00:15:02.178 { 00:15:02.178 "name": "spare", 00:15:02.178 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:02.178 "is_configured": true, 00:15:02.178 "data_offset": 2048, 00:15:02.178 "data_size": 63488 00:15:02.178 }, 00:15:02.178 { 00:15:02.178 "name": "BaseBdev2", 00:15:02.178 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:02.178 "is_configured": true, 00:15:02.178 "data_offset": 2048, 00:15:02.178 "data_size": 63488 00:15:02.178 }, 00:15:02.178 { 00:15:02.178 "name": "BaseBdev3", 00:15:02.178 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:02.178 "is_configured": true, 00:15:02.178 "data_offset": 2048, 00:15:02.178 "data_size": 63488 00:15:02.178 }, 00:15:02.178 { 00:15:02.178 "name": "BaseBdev4", 00:15:02.178 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:02.178 "is_configured": true, 00:15:02.178 "data_offset": 2048, 00:15:02.178 "data_size": 63488 00:15:02.178 } 00:15:02.178 ] 00:15:02.178 }' 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:02.178 14:14:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.130 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.131 "name": "raid_bdev1", 00:15:03.131 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:03.131 "strip_size_kb": 64, 00:15:03.131 "state": "online", 00:15:03.131 "raid_level": "raid5f", 00:15:03.131 "superblock": true, 00:15:03.131 "num_base_bdevs": 4, 00:15:03.131 "num_base_bdevs_discovered": 4, 00:15:03.131 "num_base_bdevs_operational": 4, 00:15:03.131 "process": { 00:15:03.131 "type": "rebuild", 00:15:03.131 "target": "spare", 00:15:03.131 "progress": { 00:15:03.131 "blocks": 65280, 00:15:03.131 "percent": 34 00:15:03.131 } 00:15:03.131 }, 00:15:03.131 "base_bdevs_list": [ 00:15:03.131 { 00:15:03.131 "name": "spare", 00:15:03.131 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:03.131 "is_configured": true, 00:15:03.131 "data_offset": 2048, 00:15:03.131 "data_size": 63488 00:15:03.131 }, 00:15:03.131 { 00:15:03.131 "name": "BaseBdev2", 00:15:03.131 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:03.131 "is_configured": true, 00:15:03.131 "data_offset": 2048, 00:15:03.131 "data_size": 63488 00:15:03.131 }, 00:15:03.131 { 00:15:03.131 "name": "BaseBdev3", 00:15:03.131 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:03.131 "is_configured": true, 00:15:03.131 "data_offset": 2048, 00:15:03.131 "data_size": 63488 00:15:03.131 }, 00:15:03.131 { 00:15:03.131 "name": "BaseBdev4", 00:15:03.131 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:03.131 "is_configured": true, 00:15:03.131 "data_offset": 2048, 00:15:03.131 "data_size": 63488 00:15:03.131 } 00:15:03.131 ] 00:15:03.131 }' 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:03.131 14:14:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.068 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.327 "name": "raid_bdev1", 00:15:04.327 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:04.327 "strip_size_kb": 64, 00:15:04.327 "state": "online", 00:15:04.327 "raid_level": "raid5f", 00:15:04.327 "superblock": true, 00:15:04.327 "num_base_bdevs": 4, 00:15:04.327 "num_base_bdevs_discovered": 4, 00:15:04.327 "num_base_bdevs_operational": 4, 00:15:04.327 "process": { 00:15:04.327 "type": "rebuild", 00:15:04.327 "target": "spare", 00:15:04.327 "progress": { 00:15:04.327 "blocks": 86400, 00:15:04.327 "percent": 45 00:15:04.327 } 00:15:04.327 }, 00:15:04.327 "base_bdevs_list": [ 00:15:04.327 { 00:15:04.327 "name": "spare", 00:15:04.327 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:04.327 "is_configured": true, 00:15:04.327 "data_offset": 2048, 00:15:04.327 "data_size": 63488 00:15:04.327 }, 00:15:04.327 { 00:15:04.327 "name": "BaseBdev2", 00:15:04.327 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:04.327 "is_configured": true, 00:15:04.327 "data_offset": 2048, 00:15:04.327 "data_size": 63488 00:15:04.327 }, 00:15:04.327 { 00:15:04.327 "name": "BaseBdev3", 00:15:04.327 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:04.327 "is_configured": true, 00:15:04.327 "data_offset": 2048, 00:15:04.327 "data_size": 63488 00:15:04.327 }, 00:15:04.327 { 00:15:04.327 "name": "BaseBdev4", 00:15:04.327 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:04.327 "is_configured": true, 00:15:04.327 "data_offset": 2048, 00:15:04.327 "data_size": 63488 00:15:04.327 } 00:15:04.327 ] 00:15:04.327 }' 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:04.327 14:14:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:05.265 "name": "raid_bdev1", 00:15:05.265 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:05.265 "strip_size_kb": 64, 00:15:05.265 "state": "online", 00:15:05.265 "raid_level": "raid5f", 00:15:05.265 "superblock": true, 00:15:05.265 "num_base_bdevs": 4, 00:15:05.265 "num_base_bdevs_discovered": 4, 00:15:05.265 "num_base_bdevs_operational": 4, 00:15:05.265 "process": { 00:15:05.265 "type": "rebuild", 00:15:05.265 "target": "spare", 00:15:05.265 "progress": { 00:15:05.265 "blocks": 109440, 00:15:05.265 "percent": 57 00:15:05.265 } 00:15:05.265 }, 00:15:05.265 "base_bdevs_list": [ 00:15:05.265 { 00:15:05.265 "name": "spare", 00:15:05.265 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:05.265 "is_configured": true, 00:15:05.265 "data_offset": 2048, 00:15:05.265 "data_size": 63488 00:15:05.265 }, 00:15:05.265 { 00:15:05.265 "name": "BaseBdev2", 00:15:05.265 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:05.265 "is_configured": true, 00:15:05.265 "data_offset": 2048, 00:15:05.265 "data_size": 63488 00:15:05.265 }, 00:15:05.265 { 00:15:05.265 "name": "BaseBdev3", 00:15:05.265 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:05.265 "is_configured": true, 00:15:05.265 "data_offset": 2048, 00:15:05.265 "data_size": 63488 00:15:05.265 }, 00:15:05.265 { 00:15:05.265 "name": "BaseBdev4", 00:15:05.265 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:05.265 "is_configured": true, 00:15:05.265 "data_offset": 2048, 00:15:05.265 "data_size": 63488 00:15:05.265 } 00:15:05.265 ] 00:15:05.265 }' 00:15:05.265 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:05.524 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:05.524 14:14:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:05.524 14:14:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:05.524 14:14:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.462 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.463 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.463 "name": "raid_bdev1", 00:15:06.463 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:06.463 "strip_size_kb": 64, 00:15:06.463 "state": "online", 00:15:06.463 "raid_level": "raid5f", 00:15:06.463 "superblock": true, 00:15:06.463 "num_base_bdevs": 4, 00:15:06.463 "num_base_bdevs_discovered": 4, 00:15:06.463 "num_base_bdevs_operational": 4, 00:15:06.463 "process": { 00:15:06.463 "type": "rebuild", 00:15:06.463 "target": "spare", 00:15:06.463 "progress": { 00:15:06.463 "blocks": 130560, 00:15:06.463 "percent": 68 00:15:06.463 } 00:15:06.463 }, 00:15:06.463 "base_bdevs_list": [ 00:15:06.463 { 00:15:06.463 "name": "spare", 00:15:06.463 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:06.463 "is_configured": true, 00:15:06.463 "data_offset": 2048, 00:15:06.463 "data_size": 63488 00:15:06.463 }, 00:15:06.463 { 00:15:06.463 "name": "BaseBdev2", 00:15:06.463 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:06.463 "is_configured": true, 00:15:06.463 "data_offset": 2048, 00:15:06.463 "data_size": 63488 00:15:06.463 }, 00:15:06.463 { 00:15:06.463 "name": "BaseBdev3", 00:15:06.463 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:06.463 "is_configured": true, 00:15:06.463 "data_offset": 2048, 00:15:06.463 "data_size": 63488 00:15:06.463 }, 00:15:06.463 { 00:15:06.463 "name": "BaseBdev4", 00:15:06.463 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:06.463 "is_configured": true, 00:15:06.463 "data_offset": 2048, 00:15:06.463 "data_size": 63488 00:15:06.463 } 00:15:06.463 ] 00:15:06.463 }' 00:15:06.463 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.463 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:06.463 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.722 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:06.722 14:14:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:07.660 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.661 "name": "raid_bdev1", 00:15:07.661 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:07.661 "strip_size_kb": 64, 00:15:07.661 "state": "online", 00:15:07.661 "raid_level": "raid5f", 00:15:07.661 "superblock": true, 00:15:07.661 "num_base_bdevs": 4, 00:15:07.661 "num_base_bdevs_discovered": 4, 00:15:07.661 "num_base_bdevs_operational": 4, 00:15:07.661 "process": { 00:15:07.661 "type": "rebuild", 00:15:07.661 "target": "spare", 00:15:07.661 "progress": { 00:15:07.661 "blocks": 153600, 00:15:07.661 "percent": 80 00:15:07.661 } 00:15:07.661 }, 00:15:07.661 "base_bdevs_list": [ 00:15:07.661 { 00:15:07.661 "name": "spare", 00:15:07.661 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:07.661 "is_configured": true, 00:15:07.661 "data_offset": 2048, 00:15:07.661 "data_size": 63488 00:15:07.661 }, 00:15:07.661 { 00:15:07.661 "name": "BaseBdev2", 00:15:07.661 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:07.661 "is_configured": true, 00:15:07.661 "data_offset": 2048, 00:15:07.661 "data_size": 63488 00:15:07.661 }, 00:15:07.661 { 00:15:07.661 "name": "BaseBdev3", 00:15:07.661 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:07.661 "is_configured": true, 00:15:07.661 "data_offset": 2048, 00:15:07.661 "data_size": 63488 00:15:07.661 }, 00:15:07.661 { 00:15:07.661 "name": "BaseBdev4", 00:15:07.661 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:07.661 "is_configured": true, 00:15:07.661 "data_offset": 2048, 00:15:07.661 "data_size": 63488 00:15:07.661 } 00:15:07.661 ] 00:15:07.661 }' 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:07.661 14:14:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.041 "name": "raid_bdev1", 00:15:09.041 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:09.041 "strip_size_kb": 64, 00:15:09.041 "state": "online", 00:15:09.041 "raid_level": "raid5f", 00:15:09.041 "superblock": true, 00:15:09.041 "num_base_bdevs": 4, 00:15:09.041 "num_base_bdevs_discovered": 4, 00:15:09.041 "num_base_bdevs_operational": 4, 00:15:09.041 "process": { 00:15:09.041 "type": "rebuild", 00:15:09.041 "target": "spare", 00:15:09.041 "progress": { 00:15:09.041 "blocks": 174720, 00:15:09.041 "percent": 91 00:15:09.041 } 00:15:09.041 }, 00:15:09.041 "base_bdevs_list": [ 00:15:09.041 { 00:15:09.041 "name": "spare", 00:15:09.041 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:09.041 "is_configured": true, 00:15:09.041 "data_offset": 2048, 00:15:09.041 "data_size": 63488 00:15:09.041 }, 00:15:09.041 { 00:15:09.041 "name": "BaseBdev2", 00:15:09.041 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:09.041 "is_configured": true, 00:15:09.041 "data_offset": 2048, 00:15:09.041 "data_size": 63488 00:15:09.041 }, 00:15:09.041 { 00:15:09.041 "name": "BaseBdev3", 00:15:09.041 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:09.041 "is_configured": true, 00:15:09.041 "data_offset": 2048, 00:15:09.041 "data_size": 63488 00:15:09.041 }, 00:15:09.041 { 00:15:09.041 "name": "BaseBdev4", 00:15:09.041 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:09.041 "is_configured": true, 00:15:09.041 "data_offset": 2048, 00:15:09.041 "data_size": 63488 00:15:09.041 } 00:15:09.041 ] 00:15:09.041 }' 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:09.041 14:14:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:09.608 [2024-09-30 14:14:14.161720] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:09.608 [2024-09-30 14:14:14.161842] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:09.608 [2024-09-30 14:14:14.161967] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.868 "name": "raid_bdev1", 00:15:09.868 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:09.868 "strip_size_kb": 64, 00:15:09.868 "state": "online", 00:15:09.868 "raid_level": "raid5f", 00:15:09.868 "superblock": true, 00:15:09.868 "num_base_bdevs": 4, 00:15:09.868 "num_base_bdevs_discovered": 4, 00:15:09.868 "num_base_bdevs_operational": 4, 00:15:09.868 "base_bdevs_list": [ 00:15:09.868 { 00:15:09.868 "name": "spare", 00:15:09.868 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:09.868 "is_configured": true, 00:15:09.868 "data_offset": 2048, 00:15:09.868 "data_size": 63488 00:15:09.868 }, 00:15:09.868 { 00:15:09.868 "name": "BaseBdev2", 00:15:09.868 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:09.868 "is_configured": true, 00:15:09.868 "data_offset": 2048, 00:15:09.868 "data_size": 63488 00:15:09.868 }, 00:15:09.868 { 00:15:09.868 "name": "BaseBdev3", 00:15:09.868 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:09.868 "is_configured": true, 00:15:09.868 "data_offset": 2048, 00:15:09.868 "data_size": 63488 00:15:09.868 }, 00:15:09.868 { 00:15:09.868 "name": "BaseBdev4", 00:15:09.868 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:09.868 "is_configured": true, 00:15:09.868 "data_offset": 2048, 00:15:09.868 "data_size": 63488 00:15:09.868 } 00:15:09.868 ] 00:15:09.868 }' 00:15:09.868 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:10.128 "name": "raid_bdev1", 00:15:10.128 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:10.128 "strip_size_kb": 64, 00:15:10.128 "state": "online", 00:15:10.128 "raid_level": "raid5f", 00:15:10.128 "superblock": true, 00:15:10.128 "num_base_bdevs": 4, 00:15:10.128 "num_base_bdevs_discovered": 4, 00:15:10.128 "num_base_bdevs_operational": 4, 00:15:10.128 "base_bdevs_list": [ 00:15:10.128 { 00:15:10.128 "name": "spare", 00:15:10.128 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:10.128 "is_configured": true, 00:15:10.128 "data_offset": 2048, 00:15:10.128 "data_size": 63488 00:15:10.128 }, 00:15:10.128 { 00:15:10.128 "name": "BaseBdev2", 00:15:10.128 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:10.128 "is_configured": true, 00:15:10.128 "data_offset": 2048, 00:15:10.128 "data_size": 63488 00:15:10.128 }, 00:15:10.128 { 00:15:10.128 "name": "BaseBdev3", 00:15:10.128 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:10.128 "is_configured": true, 00:15:10.128 "data_offset": 2048, 00:15:10.128 "data_size": 63488 00:15:10.128 }, 00:15:10.128 { 00:15:10.128 "name": "BaseBdev4", 00:15:10.128 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:10.128 "is_configured": true, 00:15:10.128 "data_offset": 2048, 00:15:10.128 "data_size": 63488 00:15:10.128 } 00:15:10.128 ] 00:15:10.128 }' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.128 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.129 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.388 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.388 "name": "raid_bdev1", 00:15:10.388 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:10.388 "strip_size_kb": 64, 00:15:10.388 "state": "online", 00:15:10.388 "raid_level": "raid5f", 00:15:10.388 "superblock": true, 00:15:10.388 "num_base_bdevs": 4, 00:15:10.388 "num_base_bdevs_discovered": 4, 00:15:10.388 "num_base_bdevs_operational": 4, 00:15:10.388 "base_bdevs_list": [ 00:15:10.388 { 00:15:10.388 "name": "spare", 00:15:10.388 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:10.388 "is_configured": true, 00:15:10.388 "data_offset": 2048, 00:15:10.388 "data_size": 63488 00:15:10.388 }, 00:15:10.388 { 00:15:10.388 "name": "BaseBdev2", 00:15:10.388 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:10.388 "is_configured": true, 00:15:10.388 "data_offset": 2048, 00:15:10.388 "data_size": 63488 00:15:10.388 }, 00:15:10.388 { 00:15:10.388 "name": "BaseBdev3", 00:15:10.388 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:10.388 "is_configured": true, 00:15:10.388 "data_offset": 2048, 00:15:10.388 "data_size": 63488 00:15:10.388 }, 00:15:10.388 { 00:15:10.388 "name": "BaseBdev4", 00:15:10.388 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:10.388 "is_configured": true, 00:15:10.388 "data_offset": 2048, 00:15:10.388 "data_size": 63488 00:15:10.388 } 00:15:10.388 ] 00:15:10.388 }' 00:15:10.388 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.388 14:14:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.648 [2024-09-30 14:14:15.181127] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:10.648 [2024-09-30 14:14:15.181156] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:10.648 [2024-09-30 14:14:15.181218] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:10.648 [2024-09-30 14:14:15.181297] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:10.648 [2024-09-30 14:14:15.181308] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:10.648 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:10.907 /dev/nbd0 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:10.907 1+0 records in 00:15:10.907 1+0 records out 00:15:10.907 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00054078 s, 7.6 MB/s 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:10.907 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:11.165 /dev/nbd1 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:11.165 1+0 records in 00:15:11.165 1+0 records out 00:15:11.165 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000501737 s, 8.2 MB/s 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:11.165 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:11.424 14:14:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:11.424 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.684 [2024-09-30 14:14:16.282422] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:11.684 [2024-09-30 14:14:16.282493] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:11.684 [2024-09-30 14:14:16.282518] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:15:11.684 [2024-09-30 14:14:16.282528] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:11.684 [2024-09-30 14:14:16.284622] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:11.684 [2024-09-30 14:14:16.284758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:11.684 [2024-09-30 14:14:16.284844] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:11.684 [2024-09-30 14:14:16.284891] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:11.684 [2024-09-30 14:14:16.285000] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:11.684 [2024-09-30 14:14:16.285088] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:11.684 [2024-09-30 14:14:16.285161] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:11.684 spare 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.684 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.943 [2024-09-30 14:14:16.385047] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:15:11.943 [2024-09-30 14:14:16.385083] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:11.943 [2024-09-30 14:14:16.385329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000492a0 00:15:11.943 [2024-09-30 14:14:16.385769] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:15:11.943 [2024-09-30 14:14:16.385783] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:15:11.943 [2024-09-30 14:14:16.385911] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.944 "name": "raid_bdev1", 00:15:11.944 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:11.944 "strip_size_kb": 64, 00:15:11.944 "state": "online", 00:15:11.944 "raid_level": "raid5f", 00:15:11.944 "superblock": true, 00:15:11.944 "num_base_bdevs": 4, 00:15:11.944 "num_base_bdevs_discovered": 4, 00:15:11.944 "num_base_bdevs_operational": 4, 00:15:11.944 "base_bdevs_list": [ 00:15:11.944 { 00:15:11.944 "name": "spare", 00:15:11.944 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:11.944 "is_configured": true, 00:15:11.944 "data_offset": 2048, 00:15:11.944 "data_size": 63488 00:15:11.944 }, 00:15:11.944 { 00:15:11.944 "name": "BaseBdev2", 00:15:11.944 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:11.944 "is_configured": true, 00:15:11.944 "data_offset": 2048, 00:15:11.944 "data_size": 63488 00:15:11.944 }, 00:15:11.944 { 00:15:11.944 "name": "BaseBdev3", 00:15:11.944 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:11.944 "is_configured": true, 00:15:11.944 "data_offset": 2048, 00:15:11.944 "data_size": 63488 00:15:11.944 }, 00:15:11.944 { 00:15:11.944 "name": "BaseBdev4", 00:15:11.944 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:11.944 "is_configured": true, 00:15:11.944 "data_offset": 2048, 00:15:11.944 "data_size": 63488 00:15:11.944 } 00:15:11.944 ] 00:15:11.944 }' 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.944 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.203 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:12.463 "name": "raid_bdev1", 00:15:12.463 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:12.463 "strip_size_kb": 64, 00:15:12.463 "state": "online", 00:15:12.463 "raid_level": "raid5f", 00:15:12.463 "superblock": true, 00:15:12.463 "num_base_bdevs": 4, 00:15:12.463 "num_base_bdevs_discovered": 4, 00:15:12.463 "num_base_bdevs_operational": 4, 00:15:12.463 "base_bdevs_list": [ 00:15:12.463 { 00:15:12.463 "name": "spare", 00:15:12.463 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev2", 00:15:12.463 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev3", 00:15:12.463 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev4", 00:15:12.463 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 } 00:15:12.463 ] 00:15:12.463 }' 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.463 14:14:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.463 [2024-09-30 14:14:17.022343] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:12.463 "name": "raid_bdev1", 00:15:12.463 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:12.463 "strip_size_kb": 64, 00:15:12.463 "state": "online", 00:15:12.463 "raid_level": "raid5f", 00:15:12.463 "superblock": true, 00:15:12.463 "num_base_bdevs": 4, 00:15:12.463 "num_base_bdevs_discovered": 3, 00:15:12.463 "num_base_bdevs_operational": 3, 00:15:12.463 "base_bdevs_list": [ 00:15:12.463 { 00:15:12.463 "name": null, 00:15:12.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:12.463 "is_configured": false, 00:15:12.463 "data_offset": 0, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev2", 00:15:12.463 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev3", 00:15:12.463 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 }, 00:15:12.463 { 00:15:12.463 "name": "BaseBdev4", 00:15:12.463 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:12.463 "is_configured": true, 00:15:12.463 "data_offset": 2048, 00:15:12.463 "data_size": 63488 00:15:12.463 } 00:15:12.463 ] 00:15:12.463 }' 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:12.463 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.032 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:13.032 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.032 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.032 [2024-09-30 14:14:17.449625] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:13.032 [2024-09-30 14:14:17.449750] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:13.032 [2024-09-30 14:14:17.449768] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:13.032 [2024-09-30 14:14:17.449801] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:13.032 [2024-09-30 14:14:17.452809] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049370 00:15:13.032 [2024-09-30 14:14:17.454888] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:13.032 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.032 14:14:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.970 "name": "raid_bdev1", 00:15:13.970 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:13.970 "strip_size_kb": 64, 00:15:13.970 "state": "online", 00:15:13.970 "raid_level": "raid5f", 00:15:13.970 "superblock": true, 00:15:13.970 "num_base_bdevs": 4, 00:15:13.970 "num_base_bdevs_discovered": 4, 00:15:13.970 "num_base_bdevs_operational": 4, 00:15:13.970 "process": { 00:15:13.970 "type": "rebuild", 00:15:13.970 "target": "spare", 00:15:13.970 "progress": { 00:15:13.970 "blocks": 19200, 00:15:13.970 "percent": 10 00:15:13.970 } 00:15:13.970 }, 00:15:13.970 "base_bdevs_list": [ 00:15:13.970 { 00:15:13.970 "name": "spare", 00:15:13.970 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:13.970 "is_configured": true, 00:15:13.970 "data_offset": 2048, 00:15:13.970 "data_size": 63488 00:15:13.970 }, 00:15:13.970 { 00:15:13.970 "name": "BaseBdev2", 00:15:13.970 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:13.970 "is_configured": true, 00:15:13.970 "data_offset": 2048, 00:15:13.970 "data_size": 63488 00:15:13.970 }, 00:15:13.970 { 00:15:13.970 "name": "BaseBdev3", 00:15:13.970 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:13.970 "is_configured": true, 00:15:13.970 "data_offset": 2048, 00:15:13.970 "data_size": 63488 00:15:13.970 }, 00:15:13.970 { 00:15:13.970 "name": "BaseBdev4", 00:15:13.970 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:13.970 "is_configured": true, 00:15:13.970 "data_offset": 2048, 00:15:13.970 "data_size": 63488 00:15:13.970 } 00:15:13.970 ] 00:15:13.970 }' 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.970 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.970 [2024-09-30 14:14:18.597708] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:14.230 [2024-09-30 14:14:18.659803] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:14.230 [2024-09-30 14:14:18.659906] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.230 [2024-09-30 14:14:18.659939] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:14.230 [2024-09-30 14:14:18.659964] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.230 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.230 "name": "raid_bdev1", 00:15:14.230 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:14.230 "strip_size_kb": 64, 00:15:14.230 "state": "online", 00:15:14.230 "raid_level": "raid5f", 00:15:14.230 "superblock": true, 00:15:14.230 "num_base_bdevs": 4, 00:15:14.230 "num_base_bdevs_discovered": 3, 00:15:14.230 "num_base_bdevs_operational": 3, 00:15:14.230 "base_bdevs_list": [ 00:15:14.231 { 00:15:14.231 "name": null, 00:15:14.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:14.231 "is_configured": false, 00:15:14.231 "data_offset": 0, 00:15:14.231 "data_size": 63488 00:15:14.231 }, 00:15:14.231 { 00:15:14.231 "name": "BaseBdev2", 00:15:14.231 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:14.231 "is_configured": true, 00:15:14.231 "data_offset": 2048, 00:15:14.231 "data_size": 63488 00:15:14.231 }, 00:15:14.231 { 00:15:14.231 "name": "BaseBdev3", 00:15:14.231 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:14.231 "is_configured": true, 00:15:14.231 "data_offset": 2048, 00:15:14.231 "data_size": 63488 00:15:14.231 }, 00:15:14.231 { 00:15:14.231 "name": "BaseBdev4", 00:15:14.231 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:14.231 "is_configured": true, 00:15:14.231 "data_offset": 2048, 00:15:14.231 "data_size": 63488 00:15:14.231 } 00:15:14.231 ] 00:15:14.231 }' 00:15:14.231 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.231 14:14:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:14.490 14:14:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:14.490 14:14:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.490 14:14:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:14.490 [2024-09-30 14:14:19.115753] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:14.490 [2024-09-30 14:14:19.115859] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.490 [2024-09-30 14:14:19.115899] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:15:14.490 [2024-09-30 14:14:19.115954] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.490 [2024-09-30 14:14:19.116355] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.490 [2024-09-30 14:14:19.116413] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:14.490 [2024-09-30 14:14:19.116513] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:14.490 [2024-09-30 14:14:19.116556] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:14.490 [2024-09-30 14:14:19.116611] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:14.490 [2024-09-30 14:14:19.116686] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:14.490 [2024-09-30 14:14:19.119140] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049440 00:15:14.490 [2024-09-30 14:14:19.121291] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:14.490 spare 00:15:14.490 14:14:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.490 14:14:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.870 "name": "raid_bdev1", 00:15:15.870 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:15.870 "strip_size_kb": 64, 00:15:15.870 "state": "online", 00:15:15.870 "raid_level": "raid5f", 00:15:15.870 "superblock": true, 00:15:15.870 "num_base_bdevs": 4, 00:15:15.870 "num_base_bdevs_discovered": 4, 00:15:15.870 "num_base_bdevs_operational": 4, 00:15:15.870 "process": { 00:15:15.870 "type": "rebuild", 00:15:15.870 "target": "spare", 00:15:15.870 "progress": { 00:15:15.870 "blocks": 19200, 00:15:15.870 "percent": 10 00:15:15.870 } 00:15:15.870 }, 00:15:15.870 "base_bdevs_list": [ 00:15:15.870 { 00:15:15.870 "name": "spare", 00:15:15.870 "uuid": "169f3fe4-8991-5813-8a4f-35f98d487a8e", 00:15:15.870 "is_configured": true, 00:15:15.870 "data_offset": 2048, 00:15:15.870 "data_size": 63488 00:15:15.870 }, 00:15:15.870 { 00:15:15.870 "name": "BaseBdev2", 00:15:15.870 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:15.870 "is_configured": true, 00:15:15.870 "data_offset": 2048, 00:15:15.870 "data_size": 63488 00:15:15.870 }, 00:15:15.870 { 00:15:15.870 "name": "BaseBdev3", 00:15:15.870 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:15.870 "is_configured": true, 00:15:15.870 "data_offset": 2048, 00:15:15.870 "data_size": 63488 00:15:15.870 }, 00:15:15.870 { 00:15:15.870 "name": "BaseBdev4", 00:15:15.870 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:15.870 "is_configured": true, 00:15:15.870 "data_offset": 2048, 00:15:15.870 "data_size": 63488 00:15:15.870 } 00:15:15.870 ] 00:15:15.870 }' 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.870 [2024-09-30 14:14:20.288236] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:15.870 [2024-09-30 14:14:20.326093] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:15.870 [2024-09-30 14:14:20.326177] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:15.870 [2024-09-30 14:14:20.326197] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:15.870 [2024-09-30 14:14:20.326204] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.870 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.871 "name": "raid_bdev1", 00:15:15.871 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:15.871 "strip_size_kb": 64, 00:15:15.871 "state": "online", 00:15:15.871 "raid_level": "raid5f", 00:15:15.871 "superblock": true, 00:15:15.871 "num_base_bdevs": 4, 00:15:15.871 "num_base_bdevs_discovered": 3, 00:15:15.871 "num_base_bdevs_operational": 3, 00:15:15.871 "base_bdevs_list": [ 00:15:15.871 { 00:15:15.871 "name": null, 00:15:15.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.871 "is_configured": false, 00:15:15.871 "data_offset": 0, 00:15:15.871 "data_size": 63488 00:15:15.871 }, 00:15:15.871 { 00:15:15.871 "name": "BaseBdev2", 00:15:15.871 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:15.871 "is_configured": true, 00:15:15.871 "data_offset": 2048, 00:15:15.871 "data_size": 63488 00:15:15.871 }, 00:15:15.871 { 00:15:15.871 "name": "BaseBdev3", 00:15:15.871 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:15.871 "is_configured": true, 00:15:15.871 "data_offset": 2048, 00:15:15.871 "data_size": 63488 00:15:15.871 }, 00:15:15.871 { 00:15:15.871 "name": "BaseBdev4", 00:15:15.871 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:15.871 "is_configured": true, 00:15:15.871 "data_offset": 2048, 00:15:15.871 "data_size": 63488 00:15:15.871 } 00:15:15.871 ] 00:15:15.871 }' 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.871 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.130 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:16.390 "name": "raid_bdev1", 00:15:16.390 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:16.390 "strip_size_kb": 64, 00:15:16.390 "state": "online", 00:15:16.390 "raid_level": "raid5f", 00:15:16.390 "superblock": true, 00:15:16.390 "num_base_bdevs": 4, 00:15:16.390 "num_base_bdevs_discovered": 3, 00:15:16.390 "num_base_bdevs_operational": 3, 00:15:16.390 "base_bdevs_list": [ 00:15:16.390 { 00:15:16.390 "name": null, 00:15:16.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.390 "is_configured": false, 00:15:16.390 "data_offset": 0, 00:15:16.390 "data_size": 63488 00:15:16.390 }, 00:15:16.390 { 00:15:16.390 "name": "BaseBdev2", 00:15:16.390 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:16.390 "is_configured": true, 00:15:16.390 "data_offset": 2048, 00:15:16.390 "data_size": 63488 00:15:16.390 }, 00:15:16.390 { 00:15:16.390 "name": "BaseBdev3", 00:15:16.390 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:16.390 "is_configured": true, 00:15:16.390 "data_offset": 2048, 00:15:16.390 "data_size": 63488 00:15:16.390 }, 00:15:16.390 { 00:15:16.390 "name": "BaseBdev4", 00:15:16.390 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:16.390 "is_configured": true, 00:15:16.390 "data_offset": 2048, 00:15:16.390 "data_size": 63488 00:15:16.390 } 00:15:16.390 ] 00:15:16.390 }' 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.390 [2024-09-30 14:14:20.917757] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:16.390 [2024-09-30 14:14:20.917801] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.390 [2024-09-30 14:14:20.917822] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:15:16.390 [2024-09-30 14:14:20.917831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.390 [2024-09-30 14:14:20.918195] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.390 [2024-09-30 14:14:20.918211] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:16.390 [2024-09-30 14:14:20.918273] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:16.390 [2024-09-30 14:14:20.918295] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:16.390 [2024-09-30 14:14:20.918307] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:16.390 [2024-09-30 14:14:20.918321] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:16.390 BaseBdev1 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.390 14:14:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:17.329 "name": "raid_bdev1", 00:15:17.329 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:17.329 "strip_size_kb": 64, 00:15:17.329 "state": "online", 00:15:17.329 "raid_level": "raid5f", 00:15:17.329 "superblock": true, 00:15:17.329 "num_base_bdevs": 4, 00:15:17.329 "num_base_bdevs_discovered": 3, 00:15:17.329 "num_base_bdevs_operational": 3, 00:15:17.329 "base_bdevs_list": [ 00:15:17.329 { 00:15:17.329 "name": null, 00:15:17.329 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:17.329 "is_configured": false, 00:15:17.329 "data_offset": 0, 00:15:17.329 "data_size": 63488 00:15:17.329 }, 00:15:17.329 { 00:15:17.329 "name": "BaseBdev2", 00:15:17.329 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:17.329 "is_configured": true, 00:15:17.329 "data_offset": 2048, 00:15:17.329 "data_size": 63488 00:15:17.329 }, 00:15:17.329 { 00:15:17.329 "name": "BaseBdev3", 00:15:17.329 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:17.329 "is_configured": true, 00:15:17.329 "data_offset": 2048, 00:15:17.329 "data_size": 63488 00:15:17.329 }, 00:15:17.329 { 00:15:17.329 "name": "BaseBdev4", 00:15:17.329 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:17.329 "is_configured": true, 00:15:17.329 "data_offset": 2048, 00:15:17.329 "data_size": 63488 00:15:17.329 } 00:15:17.329 ] 00:15:17.329 }' 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:17.329 14:14:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:17.898 "name": "raid_bdev1", 00:15:17.898 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:17.898 "strip_size_kb": 64, 00:15:17.898 "state": "online", 00:15:17.898 "raid_level": "raid5f", 00:15:17.898 "superblock": true, 00:15:17.898 "num_base_bdevs": 4, 00:15:17.898 "num_base_bdevs_discovered": 3, 00:15:17.898 "num_base_bdevs_operational": 3, 00:15:17.898 "base_bdevs_list": [ 00:15:17.898 { 00:15:17.898 "name": null, 00:15:17.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:17.898 "is_configured": false, 00:15:17.898 "data_offset": 0, 00:15:17.898 "data_size": 63488 00:15:17.898 }, 00:15:17.898 { 00:15:17.898 "name": "BaseBdev2", 00:15:17.898 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:17.898 "is_configured": true, 00:15:17.898 "data_offset": 2048, 00:15:17.898 "data_size": 63488 00:15:17.898 }, 00:15:17.898 { 00:15:17.898 "name": "BaseBdev3", 00:15:17.898 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:17.898 "is_configured": true, 00:15:17.898 "data_offset": 2048, 00:15:17.898 "data_size": 63488 00:15:17.898 }, 00:15:17.898 { 00:15:17.898 "name": "BaseBdev4", 00:15:17.898 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:17.898 "is_configured": true, 00:15:17.898 "data_offset": 2048, 00:15:17.898 "data_size": 63488 00:15:17.898 } 00:15:17.898 ] 00:15:17.898 }' 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.898 [2024-09-30 14:14:22.495169] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:17.898 [2024-09-30 14:14:22.495341] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:17.898 [2024-09-30 14:14:22.495363] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:17.898 request: 00:15:17.898 { 00:15:17.898 "base_bdev": "BaseBdev1", 00:15:17.898 "raid_bdev": "raid_bdev1", 00:15:17.898 "method": "bdev_raid_add_base_bdev", 00:15:17.898 "req_id": 1 00:15:17.898 } 00:15:17.898 Got JSON-RPC error response 00:15:17.898 response: 00:15:17.898 { 00:15:17.898 "code": -22, 00:15:17.898 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:17.898 } 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:17.898 14:14:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.277 "name": "raid_bdev1", 00:15:19.277 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:19.277 "strip_size_kb": 64, 00:15:19.277 "state": "online", 00:15:19.277 "raid_level": "raid5f", 00:15:19.277 "superblock": true, 00:15:19.277 "num_base_bdevs": 4, 00:15:19.277 "num_base_bdevs_discovered": 3, 00:15:19.277 "num_base_bdevs_operational": 3, 00:15:19.277 "base_bdevs_list": [ 00:15:19.277 { 00:15:19.277 "name": null, 00:15:19.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.277 "is_configured": false, 00:15:19.277 "data_offset": 0, 00:15:19.277 "data_size": 63488 00:15:19.277 }, 00:15:19.277 { 00:15:19.277 "name": "BaseBdev2", 00:15:19.277 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:19.277 "is_configured": true, 00:15:19.277 "data_offset": 2048, 00:15:19.277 "data_size": 63488 00:15:19.277 }, 00:15:19.277 { 00:15:19.277 "name": "BaseBdev3", 00:15:19.277 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:19.277 "is_configured": true, 00:15:19.277 "data_offset": 2048, 00:15:19.277 "data_size": 63488 00:15:19.277 }, 00:15:19.277 { 00:15:19.277 "name": "BaseBdev4", 00:15:19.277 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:19.277 "is_configured": true, 00:15:19.277 "data_offset": 2048, 00:15:19.277 "data_size": 63488 00:15:19.277 } 00:15:19.277 ] 00:15:19.277 }' 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.277 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.537 14:14:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:19.537 "name": "raid_bdev1", 00:15:19.537 "uuid": "6d236ccd-ee82-4c46-a965-4e4210a51e96", 00:15:19.537 "strip_size_kb": 64, 00:15:19.537 "state": "online", 00:15:19.537 "raid_level": "raid5f", 00:15:19.537 "superblock": true, 00:15:19.537 "num_base_bdevs": 4, 00:15:19.537 "num_base_bdevs_discovered": 3, 00:15:19.537 "num_base_bdevs_operational": 3, 00:15:19.537 "base_bdevs_list": [ 00:15:19.537 { 00:15:19.537 "name": null, 00:15:19.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.537 "is_configured": false, 00:15:19.537 "data_offset": 0, 00:15:19.537 "data_size": 63488 00:15:19.537 }, 00:15:19.537 { 00:15:19.537 "name": "BaseBdev2", 00:15:19.537 "uuid": "0d7c6575-29b0-58a0-8229-00d3989d2771", 00:15:19.537 "is_configured": true, 00:15:19.537 "data_offset": 2048, 00:15:19.537 "data_size": 63488 00:15:19.537 }, 00:15:19.537 { 00:15:19.537 "name": "BaseBdev3", 00:15:19.537 "uuid": "6fc2bf70-baeb-5820-ba6a-0f103e31266e", 00:15:19.537 "is_configured": true, 00:15:19.537 "data_offset": 2048, 00:15:19.537 "data_size": 63488 00:15:19.537 }, 00:15:19.537 { 00:15:19.537 "name": "BaseBdev4", 00:15:19.537 "uuid": "22f2266e-65aa-53e3-b86c-b4dae53f1e28", 00:15:19.537 "is_configured": true, 00:15:19.537 "data_offset": 2048, 00:15:19.537 "data_size": 63488 00:15:19.537 } 00:15:19.537 ] 00:15:19.537 }' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 96543 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 96543 ']' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 96543 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96543 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:19.537 killing process with pid 96543 00:15:19.537 Received shutdown signal, test time was about 60.000000 seconds 00:15:19.537 00:15:19.537 Latency(us) 00:15:19.537 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:19.537 =================================================================================================================== 00:15:19.537 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96543' 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 96543 00:15:19.537 [2024-09-30 14:14:24.137309] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:19.537 [2024-09-30 14:14:24.137398] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:19.537 [2024-09-30 14:14:24.137458] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:19.537 [2024-09-30 14:14:24.137469] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:15:19.537 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 96543 00:15:19.537 [2024-09-30 14:14:24.186751] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:19.797 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:19.797 00:15:19.797 real 0m24.995s 00:15:19.797 user 0m31.718s 00:15:19.797 sys 0m2.996s 00:15:19.797 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:19.797 ************************************ 00:15:19.797 END TEST raid5f_rebuild_test_sb 00:15:19.797 ************************************ 00:15:19.797 14:14:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.057 14:14:24 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:20.057 14:14:24 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:20.057 14:14:24 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:15:20.057 14:14:24 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:20.057 14:14:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:20.057 ************************************ 00:15:20.057 START TEST raid_state_function_test_sb_4k 00:15:20.057 ************************************ 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=97331 00:15:20.057 Process raid pid: 97331 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97331' 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 97331 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 97331 ']' 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:20.057 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:20.057 14:14:24 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.057 [2024-09-30 14:14:24.606802] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:20.057 [2024-09-30 14:14:24.606948] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:20.317 [2024-09-30 14:14:24.745428] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:20.317 [2024-09-30 14:14:24.772471] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:20.317 [2024-09-30 14:14:24.819852] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:20.317 [2024-09-30 14:14:24.861725] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:20.317 [2024-09-30 14:14:24.861760] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.886 [2024-09-30 14:14:25.427012] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:20.886 [2024-09-30 14:14:25.427061] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:20.886 [2024-09-30 14:14:25.427074] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:20.886 [2024-09-30 14:14:25.427081] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:20.886 "name": "Existed_Raid", 00:15:20.886 "uuid": "04779457-5a47-476c-8a92-9f625a66b562", 00:15:20.886 "strip_size_kb": 0, 00:15:20.886 "state": "configuring", 00:15:20.886 "raid_level": "raid1", 00:15:20.886 "superblock": true, 00:15:20.886 "num_base_bdevs": 2, 00:15:20.886 "num_base_bdevs_discovered": 0, 00:15:20.886 "num_base_bdevs_operational": 2, 00:15:20.886 "base_bdevs_list": [ 00:15:20.886 { 00:15:20.886 "name": "BaseBdev1", 00:15:20.886 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.886 "is_configured": false, 00:15:20.886 "data_offset": 0, 00:15:20.886 "data_size": 0 00:15:20.886 }, 00:15:20.886 { 00:15:20.886 "name": "BaseBdev2", 00:15:20.886 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.886 "is_configured": false, 00:15:20.886 "data_offset": 0, 00:15:20.886 "data_size": 0 00:15:20.886 } 00:15:20.886 ] 00:15:20.886 }' 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:20.886 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.456 [2024-09-30 14:14:25.930119] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:21.456 [2024-09-30 14:14:25.930200] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.456 [2024-09-30 14:14:25.938111] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:21.456 [2024-09-30 14:14:25.938187] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:21.456 [2024-09-30 14:14:25.938213] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:21.456 [2024-09-30 14:14:25.938232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.456 [2024-09-30 14:14:25.954726] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:21.456 BaseBdev1 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:21.456 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.457 [ 00:15:21.457 { 00:15:21.457 "name": "BaseBdev1", 00:15:21.457 "aliases": [ 00:15:21.457 "90efded9-46b9-4e9b-a26f-9bcaf2e2e943" 00:15:21.457 ], 00:15:21.457 "product_name": "Malloc disk", 00:15:21.457 "block_size": 4096, 00:15:21.457 "num_blocks": 8192, 00:15:21.457 "uuid": "90efded9-46b9-4e9b-a26f-9bcaf2e2e943", 00:15:21.457 "assigned_rate_limits": { 00:15:21.457 "rw_ios_per_sec": 0, 00:15:21.457 "rw_mbytes_per_sec": 0, 00:15:21.457 "r_mbytes_per_sec": 0, 00:15:21.457 "w_mbytes_per_sec": 0 00:15:21.457 }, 00:15:21.457 "claimed": true, 00:15:21.457 "claim_type": "exclusive_write", 00:15:21.457 "zoned": false, 00:15:21.457 "supported_io_types": { 00:15:21.457 "read": true, 00:15:21.457 "write": true, 00:15:21.457 "unmap": true, 00:15:21.457 "flush": true, 00:15:21.457 "reset": true, 00:15:21.457 "nvme_admin": false, 00:15:21.457 "nvme_io": false, 00:15:21.457 "nvme_io_md": false, 00:15:21.457 "write_zeroes": true, 00:15:21.457 "zcopy": true, 00:15:21.457 "get_zone_info": false, 00:15:21.457 "zone_management": false, 00:15:21.457 "zone_append": false, 00:15:21.457 "compare": false, 00:15:21.457 "compare_and_write": false, 00:15:21.457 "abort": true, 00:15:21.457 "seek_hole": false, 00:15:21.457 "seek_data": false, 00:15:21.457 "copy": true, 00:15:21.457 "nvme_iov_md": false 00:15:21.457 }, 00:15:21.457 "memory_domains": [ 00:15:21.457 { 00:15:21.457 "dma_device_id": "system", 00:15:21.457 "dma_device_type": 1 00:15:21.457 }, 00:15:21.457 { 00:15:21.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:21.457 "dma_device_type": 2 00:15:21.457 } 00:15:21.457 ], 00:15:21.457 "driver_specific": {} 00:15:21.457 } 00:15:21.457 ] 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.457 14:14:25 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.457 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.457 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.457 "name": "Existed_Raid", 00:15:21.457 "uuid": "7767c12e-8412-45c8-858f-bfa4f5696b3d", 00:15:21.457 "strip_size_kb": 0, 00:15:21.457 "state": "configuring", 00:15:21.457 "raid_level": "raid1", 00:15:21.457 "superblock": true, 00:15:21.457 "num_base_bdevs": 2, 00:15:21.457 "num_base_bdevs_discovered": 1, 00:15:21.457 "num_base_bdevs_operational": 2, 00:15:21.457 "base_bdevs_list": [ 00:15:21.457 { 00:15:21.457 "name": "BaseBdev1", 00:15:21.457 "uuid": "90efded9-46b9-4e9b-a26f-9bcaf2e2e943", 00:15:21.457 "is_configured": true, 00:15:21.457 "data_offset": 256, 00:15:21.457 "data_size": 7936 00:15:21.457 }, 00:15:21.457 { 00:15:21.457 "name": "BaseBdev2", 00:15:21.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.457 "is_configured": false, 00:15:21.457 "data_offset": 0, 00:15:21.457 "data_size": 0 00:15:21.457 } 00:15:21.457 ] 00:15:21.457 }' 00:15:21.457 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.457 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.026 [2024-09-30 14:14:26.417977] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:22.026 [2024-09-30 14:14:26.418065] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.026 [2024-09-30 14:14:26.425986] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:22.026 [2024-09-30 14:14:26.427703] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:22.026 [2024-09-30 14:14:26.427770] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:22.026 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:22.027 "name": "Existed_Raid", 00:15:22.027 "uuid": "0cd146cf-604f-4280-a244-47ae4668449e", 00:15:22.027 "strip_size_kb": 0, 00:15:22.027 "state": "configuring", 00:15:22.027 "raid_level": "raid1", 00:15:22.027 "superblock": true, 00:15:22.027 "num_base_bdevs": 2, 00:15:22.027 "num_base_bdevs_discovered": 1, 00:15:22.027 "num_base_bdevs_operational": 2, 00:15:22.027 "base_bdevs_list": [ 00:15:22.027 { 00:15:22.027 "name": "BaseBdev1", 00:15:22.027 "uuid": "90efded9-46b9-4e9b-a26f-9bcaf2e2e943", 00:15:22.027 "is_configured": true, 00:15:22.027 "data_offset": 256, 00:15:22.027 "data_size": 7936 00:15:22.027 }, 00:15:22.027 { 00:15:22.027 "name": "BaseBdev2", 00:15:22.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.027 "is_configured": false, 00:15:22.027 "data_offset": 0, 00:15:22.027 "data_size": 0 00:15:22.027 } 00:15:22.027 ] 00:15:22.027 }' 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:22.027 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.287 [2024-09-30 14:14:26.899753] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:22.287 [2024-09-30 14:14:26.899958] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:22.287 [2024-09-30 14:14:26.899987] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:22.287 [2024-09-30 14:14:26.900310] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:22.287 [2024-09-30 14:14:26.900470] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:22.287 [2024-09-30 14:14:26.900498] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:15:22.287 BaseBdev2 00:15:22.287 [2024-09-30 14:14:26.900641] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.287 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.287 [ 00:15:22.287 { 00:15:22.287 "name": "BaseBdev2", 00:15:22.287 "aliases": [ 00:15:22.287 "a16f21ee-4897-4b2c-b4af-68df3f411bea" 00:15:22.287 ], 00:15:22.287 "product_name": "Malloc disk", 00:15:22.287 "block_size": 4096, 00:15:22.287 "num_blocks": 8192, 00:15:22.287 "uuid": "a16f21ee-4897-4b2c-b4af-68df3f411bea", 00:15:22.287 "assigned_rate_limits": { 00:15:22.287 "rw_ios_per_sec": 0, 00:15:22.287 "rw_mbytes_per_sec": 0, 00:15:22.287 "r_mbytes_per_sec": 0, 00:15:22.287 "w_mbytes_per_sec": 0 00:15:22.287 }, 00:15:22.287 "claimed": true, 00:15:22.287 "claim_type": "exclusive_write", 00:15:22.287 "zoned": false, 00:15:22.287 "supported_io_types": { 00:15:22.287 "read": true, 00:15:22.287 "write": true, 00:15:22.287 "unmap": true, 00:15:22.287 "flush": true, 00:15:22.287 "reset": true, 00:15:22.287 "nvme_admin": false, 00:15:22.287 "nvme_io": false, 00:15:22.287 "nvme_io_md": false, 00:15:22.287 "write_zeroes": true, 00:15:22.287 "zcopy": true, 00:15:22.287 "get_zone_info": false, 00:15:22.287 "zone_management": false, 00:15:22.287 "zone_append": false, 00:15:22.287 "compare": false, 00:15:22.287 "compare_and_write": false, 00:15:22.287 "abort": true, 00:15:22.287 "seek_hole": false, 00:15:22.287 "seek_data": false, 00:15:22.287 "copy": true, 00:15:22.287 "nvme_iov_md": false 00:15:22.288 }, 00:15:22.288 "memory_domains": [ 00:15:22.288 { 00:15:22.288 "dma_device_id": "system", 00:15:22.288 "dma_device_type": 1 00:15:22.288 }, 00:15:22.288 { 00:15:22.288 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:22.288 "dma_device_type": 2 00:15:22.288 } 00:15:22.288 ], 00:15:22.288 "driver_specific": {} 00:15:22.288 } 00:15:22.288 ] 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:22.288 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:22.560 "name": "Existed_Raid", 00:15:22.560 "uuid": "0cd146cf-604f-4280-a244-47ae4668449e", 00:15:22.560 "strip_size_kb": 0, 00:15:22.560 "state": "online", 00:15:22.560 "raid_level": "raid1", 00:15:22.560 "superblock": true, 00:15:22.560 "num_base_bdevs": 2, 00:15:22.560 "num_base_bdevs_discovered": 2, 00:15:22.560 "num_base_bdevs_operational": 2, 00:15:22.560 "base_bdevs_list": [ 00:15:22.560 { 00:15:22.560 "name": "BaseBdev1", 00:15:22.560 "uuid": "90efded9-46b9-4e9b-a26f-9bcaf2e2e943", 00:15:22.560 "is_configured": true, 00:15:22.560 "data_offset": 256, 00:15:22.560 "data_size": 7936 00:15:22.560 }, 00:15:22.560 { 00:15:22.560 "name": "BaseBdev2", 00:15:22.560 "uuid": "a16f21ee-4897-4b2c-b4af-68df3f411bea", 00:15:22.560 "is_configured": true, 00:15:22.560 "data_offset": 256, 00:15:22.560 "data_size": 7936 00:15:22.560 } 00:15:22.560 ] 00:15:22.560 }' 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:22.560 14:14:26 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.842 [2024-09-30 14:14:27.383350] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.842 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:22.842 "name": "Existed_Raid", 00:15:22.842 "aliases": [ 00:15:22.842 "0cd146cf-604f-4280-a244-47ae4668449e" 00:15:22.842 ], 00:15:22.842 "product_name": "Raid Volume", 00:15:22.842 "block_size": 4096, 00:15:22.842 "num_blocks": 7936, 00:15:22.842 "uuid": "0cd146cf-604f-4280-a244-47ae4668449e", 00:15:22.842 "assigned_rate_limits": { 00:15:22.842 "rw_ios_per_sec": 0, 00:15:22.842 "rw_mbytes_per_sec": 0, 00:15:22.842 "r_mbytes_per_sec": 0, 00:15:22.842 "w_mbytes_per_sec": 0 00:15:22.842 }, 00:15:22.842 "claimed": false, 00:15:22.842 "zoned": false, 00:15:22.842 "supported_io_types": { 00:15:22.842 "read": true, 00:15:22.842 "write": true, 00:15:22.842 "unmap": false, 00:15:22.842 "flush": false, 00:15:22.842 "reset": true, 00:15:22.842 "nvme_admin": false, 00:15:22.842 "nvme_io": false, 00:15:22.842 "nvme_io_md": false, 00:15:22.842 "write_zeroes": true, 00:15:22.842 "zcopy": false, 00:15:22.842 "get_zone_info": false, 00:15:22.842 "zone_management": false, 00:15:22.842 "zone_append": false, 00:15:22.842 "compare": false, 00:15:22.842 "compare_and_write": false, 00:15:22.842 "abort": false, 00:15:22.842 "seek_hole": false, 00:15:22.842 "seek_data": false, 00:15:22.842 "copy": false, 00:15:22.842 "nvme_iov_md": false 00:15:22.842 }, 00:15:22.842 "memory_domains": [ 00:15:22.842 { 00:15:22.842 "dma_device_id": "system", 00:15:22.842 "dma_device_type": 1 00:15:22.842 }, 00:15:22.842 { 00:15:22.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:22.842 "dma_device_type": 2 00:15:22.842 }, 00:15:22.843 { 00:15:22.843 "dma_device_id": "system", 00:15:22.843 "dma_device_type": 1 00:15:22.843 }, 00:15:22.843 { 00:15:22.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:22.843 "dma_device_type": 2 00:15:22.843 } 00:15:22.843 ], 00:15:22.843 "driver_specific": { 00:15:22.843 "raid": { 00:15:22.843 "uuid": "0cd146cf-604f-4280-a244-47ae4668449e", 00:15:22.843 "strip_size_kb": 0, 00:15:22.843 "state": "online", 00:15:22.843 "raid_level": "raid1", 00:15:22.843 "superblock": true, 00:15:22.843 "num_base_bdevs": 2, 00:15:22.843 "num_base_bdevs_discovered": 2, 00:15:22.843 "num_base_bdevs_operational": 2, 00:15:22.843 "base_bdevs_list": [ 00:15:22.843 { 00:15:22.843 "name": "BaseBdev1", 00:15:22.843 "uuid": "90efded9-46b9-4e9b-a26f-9bcaf2e2e943", 00:15:22.843 "is_configured": true, 00:15:22.843 "data_offset": 256, 00:15:22.843 "data_size": 7936 00:15:22.843 }, 00:15:22.843 { 00:15:22.843 "name": "BaseBdev2", 00:15:22.843 "uuid": "a16f21ee-4897-4b2c-b4af-68df3f411bea", 00:15:22.843 "is_configured": true, 00:15:22.843 "data_offset": 256, 00:15:22.843 "data_size": 7936 00:15:22.843 } 00:15:22.843 ] 00:15:22.843 } 00:15:22.843 } 00:15:22.843 }' 00:15:22.843 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:22.843 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:22.843 BaseBdev2' 00:15:22.843 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.103 [2024-09-30 14:14:27.622755] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:23.103 "name": "Existed_Raid", 00:15:23.103 "uuid": "0cd146cf-604f-4280-a244-47ae4668449e", 00:15:23.103 "strip_size_kb": 0, 00:15:23.103 "state": "online", 00:15:23.103 "raid_level": "raid1", 00:15:23.103 "superblock": true, 00:15:23.103 "num_base_bdevs": 2, 00:15:23.103 "num_base_bdevs_discovered": 1, 00:15:23.103 "num_base_bdevs_operational": 1, 00:15:23.103 "base_bdevs_list": [ 00:15:23.103 { 00:15:23.103 "name": null, 00:15:23.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:23.103 "is_configured": false, 00:15:23.103 "data_offset": 0, 00:15:23.103 "data_size": 7936 00:15:23.103 }, 00:15:23.103 { 00:15:23.103 "name": "BaseBdev2", 00:15:23.103 "uuid": "a16f21ee-4897-4b2c-b4af-68df3f411bea", 00:15:23.103 "is_configured": true, 00:15:23.103 "data_offset": 256, 00:15:23.103 "data_size": 7936 00:15:23.103 } 00:15:23.103 ] 00:15:23.103 }' 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:23.103 14:14:27 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.674 [2024-09-30 14:14:28.121051] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:23.674 [2024-09-30 14:14:28.121201] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:23.674 [2024-09-30 14:14:28.132457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:23.674 [2024-09-30 14:14:28.132593] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:23.674 [2024-09-30 14:14:28.132607] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 97331 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 97331 ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 97331 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97331 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:23.674 killing process with pid 97331 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97331' 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@969 -- # kill 97331 00:15:23.674 [2024-09-30 14:14:28.230949] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:23.674 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@974 -- # wait 97331 00:15:23.674 [2024-09-30 14:14:28.231913] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:23.935 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:15:23.935 00:15:23.935 real 0m3.981s 00:15:23.935 user 0m6.227s 00:15:23.935 sys 0m0.861s 00:15:23.935 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:23.935 ************************************ 00:15:23.935 END TEST raid_state_function_test_sb_4k 00:15:23.935 ************************************ 00:15:23.935 14:14:28 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.935 14:14:28 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:15:23.935 14:14:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:23.935 14:14:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:23.935 14:14:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:23.935 ************************************ 00:15:23.935 START TEST raid_superblock_test_4k 00:15:23.935 ************************************ 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=97572 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 97572 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@831 -- # '[' -z 97572 ']' 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:23.935 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:23.935 14:14:28 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.198 [2024-09-30 14:14:28.661035] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:24.198 [2024-09-30 14:14:28.661167] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97572 ] 00:15:24.198 [2024-09-30 14:14:28.798324] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:24.198 [2024-09-30 14:14:28.825632] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:24.461 [2024-09-30 14:14:28.873336] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:24.461 [2024-09-30 14:14:28.916121] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:24.461 [2024-09-30 14:14:28.916237] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # return 0 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.032 malloc1 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.032 [2024-09-30 14:14:29.486543] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:25.032 [2024-09-30 14:14:29.486686] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.032 [2024-09-30 14:14:29.486733] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:25.032 [2024-09-30 14:14:29.486763] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.032 [2024-09-30 14:14:29.488819] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.032 [2024-09-30 14:14:29.488888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:25.032 pt1 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:25.032 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.033 malloc2 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.033 [2024-09-30 14:14:29.531011] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:25.033 [2024-09-30 14:14:29.531164] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.033 [2024-09-30 14:14:29.531226] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:25.033 [2024-09-30 14:14:29.531279] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.033 [2024-09-30 14:14:29.534922] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.033 [2024-09-30 14:14:29.535031] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:25.033 pt2 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.033 [2024-09-30 14:14:29.543278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:25.033 [2024-09-30 14:14:29.545534] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:25.033 [2024-09-30 14:14:29.545697] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:25.033 [2024-09-30 14:14:29.545713] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:25.033 [2024-09-30 14:14:29.546007] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:25.033 [2024-09-30 14:14:29.546170] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:25.033 [2024-09-30 14:14:29.546186] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:25.033 [2024-09-30 14:14:29.546314] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.033 "name": "raid_bdev1", 00:15:25.033 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:25.033 "strip_size_kb": 0, 00:15:25.033 "state": "online", 00:15:25.033 "raid_level": "raid1", 00:15:25.033 "superblock": true, 00:15:25.033 "num_base_bdevs": 2, 00:15:25.033 "num_base_bdevs_discovered": 2, 00:15:25.033 "num_base_bdevs_operational": 2, 00:15:25.033 "base_bdevs_list": [ 00:15:25.033 { 00:15:25.033 "name": "pt1", 00:15:25.033 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:25.033 "is_configured": true, 00:15:25.033 "data_offset": 256, 00:15:25.033 "data_size": 7936 00:15:25.033 }, 00:15:25.033 { 00:15:25.033 "name": "pt2", 00:15:25.033 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:25.033 "is_configured": true, 00:15:25.033 "data_offset": 256, 00:15:25.033 "data_size": 7936 00:15:25.033 } 00:15:25.033 ] 00:15:25.033 }' 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.033 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.603 14:14:29 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.603 [2024-09-30 14:14:29.986714] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:25.603 "name": "raid_bdev1", 00:15:25.603 "aliases": [ 00:15:25.603 "022e0569-240f-432d-b1e9-ee8cb9b7e1f2" 00:15:25.603 ], 00:15:25.603 "product_name": "Raid Volume", 00:15:25.603 "block_size": 4096, 00:15:25.603 "num_blocks": 7936, 00:15:25.603 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:25.603 "assigned_rate_limits": { 00:15:25.603 "rw_ios_per_sec": 0, 00:15:25.603 "rw_mbytes_per_sec": 0, 00:15:25.603 "r_mbytes_per_sec": 0, 00:15:25.603 "w_mbytes_per_sec": 0 00:15:25.603 }, 00:15:25.603 "claimed": false, 00:15:25.603 "zoned": false, 00:15:25.603 "supported_io_types": { 00:15:25.603 "read": true, 00:15:25.603 "write": true, 00:15:25.603 "unmap": false, 00:15:25.603 "flush": false, 00:15:25.603 "reset": true, 00:15:25.603 "nvme_admin": false, 00:15:25.603 "nvme_io": false, 00:15:25.603 "nvme_io_md": false, 00:15:25.603 "write_zeroes": true, 00:15:25.603 "zcopy": false, 00:15:25.603 "get_zone_info": false, 00:15:25.603 "zone_management": false, 00:15:25.603 "zone_append": false, 00:15:25.603 "compare": false, 00:15:25.603 "compare_and_write": false, 00:15:25.603 "abort": false, 00:15:25.603 "seek_hole": false, 00:15:25.603 "seek_data": false, 00:15:25.603 "copy": false, 00:15:25.603 "nvme_iov_md": false 00:15:25.603 }, 00:15:25.603 "memory_domains": [ 00:15:25.603 { 00:15:25.603 "dma_device_id": "system", 00:15:25.603 "dma_device_type": 1 00:15:25.603 }, 00:15:25.603 { 00:15:25.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:25.603 "dma_device_type": 2 00:15:25.603 }, 00:15:25.603 { 00:15:25.603 "dma_device_id": "system", 00:15:25.603 "dma_device_type": 1 00:15:25.603 }, 00:15:25.603 { 00:15:25.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:25.603 "dma_device_type": 2 00:15:25.603 } 00:15:25.603 ], 00:15:25.603 "driver_specific": { 00:15:25.603 "raid": { 00:15:25.603 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:25.603 "strip_size_kb": 0, 00:15:25.603 "state": "online", 00:15:25.603 "raid_level": "raid1", 00:15:25.603 "superblock": true, 00:15:25.603 "num_base_bdevs": 2, 00:15:25.603 "num_base_bdevs_discovered": 2, 00:15:25.603 "num_base_bdevs_operational": 2, 00:15:25.603 "base_bdevs_list": [ 00:15:25.603 { 00:15:25.603 "name": "pt1", 00:15:25.603 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:25.603 "is_configured": true, 00:15:25.603 "data_offset": 256, 00:15:25.603 "data_size": 7936 00:15:25.603 }, 00:15:25.603 { 00:15:25.603 "name": "pt2", 00:15:25.603 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:25.603 "is_configured": true, 00:15:25.603 "data_offset": 256, 00:15:25.603 "data_size": 7936 00:15:25.603 } 00:15:25.603 ] 00:15:25.603 } 00:15:25.603 } 00:15:25.603 }' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:25.603 pt2' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:25.603 [2024-09-30 14:14:30.226216] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:25.603 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=022e0569-240f-432d-b1e9-ee8cb9b7e1f2 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z 022e0569-240f-432d-b1e9-ee8cb9b7e1f2 ']' 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.864 [2024-09-30 14:14:30.273903] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:25.864 [2024-09-30 14:14:30.273966] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:25.864 [2024-09-30 14:14:30.274029] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:25.864 [2024-09-30 14:14:30.274091] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:25.864 [2024-09-30 14:14:30.274103] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # local es=0 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:25.864 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.865 [2024-09-30 14:14:30.413681] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:25.865 [2024-09-30 14:14:30.415512] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:25.865 [2024-09-30 14:14:30.415601] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:25.865 [2024-09-30 14:14:30.415673] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:25.865 [2024-09-30 14:14:30.415710] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:25.865 [2024-09-30 14:14:30.415739] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:15:25.865 request: 00:15:25.865 { 00:15:25.865 "name": "raid_bdev1", 00:15:25.865 "raid_level": "raid1", 00:15:25.865 "base_bdevs": [ 00:15:25.865 "malloc1", 00:15:25.865 "malloc2" 00:15:25.865 ], 00:15:25.865 "superblock": false, 00:15:25.865 "method": "bdev_raid_create", 00:15:25.865 "req_id": 1 00:15:25.865 } 00:15:25.865 Got JSON-RPC error response 00:15:25.865 response: 00:15:25.865 { 00:15:25.865 "code": -17, 00:15:25.865 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:25.865 } 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # es=1 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.865 [2024-09-30 14:14:30.477577] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:25.865 [2024-09-30 14:14:30.477624] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.865 [2024-09-30 14:14:30.477637] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:25.865 [2024-09-30 14:14:30.477649] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.865 [2024-09-30 14:14:30.479714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.865 [2024-09-30 14:14:30.479753] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:25.865 [2024-09-30 14:14:30.479804] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:25.865 [2024-09-30 14:14:30.479850] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:25.865 pt1 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.865 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.125 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:26.125 "name": "raid_bdev1", 00:15:26.125 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:26.125 "strip_size_kb": 0, 00:15:26.125 "state": "configuring", 00:15:26.125 "raid_level": "raid1", 00:15:26.125 "superblock": true, 00:15:26.125 "num_base_bdevs": 2, 00:15:26.125 "num_base_bdevs_discovered": 1, 00:15:26.125 "num_base_bdevs_operational": 2, 00:15:26.125 "base_bdevs_list": [ 00:15:26.125 { 00:15:26.125 "name": "pt1", 00:15:26.125 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:26.125 "is_configured": true, 00:15:26.125 "data_offset": 256, 00:15:26.125 "data_size": 7936 00:15:26.125 }, 00:15:26.125 { 00:15:26.125 "name": null, 00:15:26.125 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:26.125 "is_configured": false, 00:15:26.125 "data_offset": 256, 00:15:26.125 "data_size": 7936 00:15:26.125 } 00:15:26.125 ] 00:15:26.125 }' 00:15:26.125 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:26.125 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.385 [2024-09-30 14:14:30.932761] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:26.385 [2024-09-30 14:14:30.932858] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:26.385 [2024-09-30 14:14:30.932890] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:26.385 [2024-09-30 14:14:30.932920] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:26.385 [2024-09-30 14:14:30.933231] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:26.385 [2024-09-30 14:14:30.933286] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:26.385 [2024-09-30 14:14:30.933360] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:26.385 [2024-09-30 14:14:30.933405] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:26.385 [2024-09-30 14:14:30.933508] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:26.385 [2024-09-30 14:14:30.933547] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:26.385 [2024-09-30 14:14:30.933762] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:26.385 [2024-09-30 14:14:30.933905] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:26.385 [2024-09-30 14:14:30.933944] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:15:26.385 [2024-09-30 14:14:30.934061] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:26.385 pt2 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:26.385 "name": "raid_bdev1", 00:15:26.385 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:26.385 "strip_size_kb": 0, 00:15:26.385 "state": "online", 00:15:26.385 "raid_level": "raid1", 00:15:26.385 "superblock": true, 00:15:26.385 "num_base_bdevs": 2, 00:15:26.385 "num_base_bdevs_discovered": 2, 00:15:26.385 "num_base_bdevs_operational": 2, 00:15:26.385 "base_bdevs_list": [ 00:15:26.385 { 00:15:26.385 "name": "pt1", 00:15:26.385 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:26.385 "is_configured": true, 00:15:26.385 "data_offset": 256, 00:15:26.385 "data_size": 7936 00:15:26.385 }, 00:15:26.385 { 00:15:26.385 "name": "pt2", 00:15:26.385 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:26.385 "is_configured": true, 00:15:26.385 "data_offset": 256, 00:15:26.385 "data_size": 7936 00:15:26.385 } 00:15:26.385 ] 00:15:26.385 }' 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:26.385 14:14:30 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:26.955 [2024-09-30 14:14:31.340382] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:26.955 "name": "raid_bdev1", 00:15:26.955 "aliases": [ 00:15:26.955 "022e0569-240f-432d-b1e9-ee8cb9b7e1f2" 00:15:26.955 ], 00:15:26.955 "product_name": "Raid Volume", 00:15:26.955 "block_size": 4096, 00:15:26.955 "num_blocks": 7936, 00:15:26.955 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:26.955 "assigned_rate_limits": { 00:15:26.955 "rw_ios_per_sec": 0, 00:15:26.955 "rw_mbytes_per_sec": 0, 00:15:26.955 "r_mbytes_per_sec": 0, 00:15:26.955 "w_mbytes_per_sec": 0 00:15:26.955 }, 00:15:26.955 "claimed": false, 00:15:26.955 "zoned": false, 00:15:26.955 "supported_io_types": { 00:15:26.955 "read": true, 00:15:26.955 "write": true, 00:15:26.955 "unmap": false, 00:15:26.955 "flush": false, 00:15:26.955 "reset": true, 00:15:26.955 "nvme_admin": false, 00:15:26.955 "nvme_io": false, 00:15:26.955 "nvme_io_md": false, 00:15:26.955 "write_zeroes": true, 00:15:26.955 "zcopy": false, 00:15:26.955 "get_zone_info": false, 00:15:26.955 "zone_management": false, 00:15:26.955 "zone_append": false, 00:15:26.955 "compare": false, 00:15:26.955 "compare_and_write": false, 00:15:26.955 "abort": false, 00:15:26.955 "seek_hole": false, 00:15:26.955 "seek_data": false, 00:15:26.955 "copy": false, 00:15:26.955 "nvme_iov_md": false 00:15:26.955 }, 00:15:26.955 "memory_domains": [ 00:15:26.955 { 00:15:26.955 "dma_device_id": "system", 00:15:26.955 "dma_device_type": 1 00:15:26.955 }, 00:15:26.955 { 00:15:26.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.955 "dma_device_type": 2 00:15:26.955 }, 00:15:26.955 { 00:15:26.955 "dma_device_id": "system", 00:15:26.955 "dma_device_type": 1 00:15:26.955 }, 00:15:26.955 { 00:15:26.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:26.955 "dma_device_type": 2 00:15:26.955 } 00:15:26.955 ], 00:15:26.955 "driver_specific": { 00:15:26.955 "raid": { 00:15:26.955 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:26.955 "strip_size_kb": 0, 00:15:26.955 "state": "online", 00:15:26.955 "raid_level": "raid1", 00:15:26.955 "superblock": true, 00:15:26.955 "num_base_bdevs": 2, 00:15:26.955 "num_base_bdevs_discovered": 2, 00:15:26.955 "num_base_bdevs_operational": 2, 00:15:26.955 "base_bdevs_list": [ 00:15:26.955 { 00:15:26.955 "name": "pt1", 00:15:26.955 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:26.955 "is_configured": true, 00:15:26.955 "data_offset": 256, 00:15:26.955 "data_size": 7936 00:15:26.955 }, 00:15:26.955 { 00:15:26.955 "name": "pt2", 00:15:26.955 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:26.955 "is_configured": true, 00:15:26.955 "data_offset": 256, 00:15:26.955 "data_size": 7936 00:15:26.955 } 00:15:26.955 ] 00:15:26.955 } 00:15:26.955 } 00:15:26.955 }' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:26.955 pt2' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.955 [2024-09-30 14:14:31.583964] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:26.955 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' 022e0569-240f-432d-b1e9-ee8cb9b7e1f2 '!=' 022e0569-240f-432d-b1e9-ee8cb9b7e1f2 ']' 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.215 [2024-09-30 14:14:31.627707] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:27.215 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.216 "name": "raid_bdev1", 00:15:27.216 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:27.216 "strip_size_kb": 0, 00:15:27.216 "state": "online", 00:15:27.216 "raid_level": "raid1", 00:15:27.216 "superblock": true, 00:15:27.216 "num_base_bdevs": 2, 00:15:27.216 "num_base_bdevs_discovered": 1, 00:15:27.216 "num_base_bdevs_operational": 1, 00:15:27.216 "base_bdevs_list": [ 00:15:27.216 { 00:15:27.216 "name": null, 00:15:27.216 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:27.216 "is_configured": false, 00:15:27.216 "data_offset": 0, 00:15:27.216 "data_size": 7936 00:15:27.216 }, 00:15:27.216 { 00:15:27.216 "name": "pt2", 00:15:27.216 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:27.216 "is_configured": true, 00:15:27.216 "data_offset": 256, 00:15:27.216 "data_size": 7936 00:15:27.216 } 00:15:27.216 ] 00:15:27.216 }' 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.216 14:14:31 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.476 [2024-09-30 14:14:32.094896] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:27.476 [2024-09-30 14:14:32.094967] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:27.476 [2024-09-30 14:14:32.095031] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:27.476 [2024-09-30 14:14:32.095077] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:27.476 [2024-09-30 14:14:32.095109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.476 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.736 [2024-09-30 14:14:32.170753] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:27.736 [2024-09-30 14:14:32.170840] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.736 [2024-09-30 14:14:32.170867] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:15:27.736 [2024-09-30 14:14:32.170896] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.736 [2024-09-30 14:14:32.172910] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.736 [2024-09-30 14:14:32.172986] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:27.736 [2024-09-30 14:14:32.173064] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:27.736 [2024-09-30 14:14:32.173108] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:27.736 [2024-09-30 14:14:32.173185] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:15:27.736 [2024-09-30 14:14:32.173223] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:27.736 [2024-09-30 14:14:32.173433] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:27.736 [2024-09-30 14:14:32.173593] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:15:27.736 [2024-09-30 14:14:32.173632] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:15:27.736 [2024-09-30 14:14:32.173754] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:27.736 pt2 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.736 "name": "raid_bdev1", 00:15:27.736 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:27.736 "strip_size_kb": 0, 00:15:27.736 "state": "online", 00:15:27.736 "raid_level": "raid1", 00:15:27.736 "superblock": true, 00:15:27.736 "num_base_bdevs": 2, 00:15:27.736 "num_base_bdevs_discovered": 1, 00:15:27.736 "num_base_bdevs_operational": 1, 00:15:27.736 "base_bdevs_list": [ 00:15:27.736 { 00:15:27.736 "name": null, 00:15:27.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:27.736 "is_configured": false, 00:15:27.736 "data_offset": 256, 00:15:27.736 "data_size": 7936 00:15:27.736 }, 00:15:27.736 { 00:15:27.736 "name": "pt2", 00:15:27.736 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:27.736 "is_configured": true, 00:15:27.736 "data_offset": 256, 00:15:27.736 "data_size": 7936 00:15:27.736 } 00:15:27.736 ] 00:15:27.736 }' 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.736 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.996 [2024-09-30 14:14:32.602011] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:27.996 [2024-09-30 14:14:32.602079] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:27.996 [2024-09-30 14:14:32.602137] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:27.996 [2024-09-30 14:14:32.602185] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:27.996 [2024-09-30 14:14:32.602214] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:27.996 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.256 [2024-09-30 14:14:32.661915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:28.256 [2024-09-30 14:14:32.661958] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:28.256 [2024-09-30 14:14:32.661979] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:15:28.256 [2024-09-30 14:14:32.661989] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:28.256 [2024-09-30 14:14:32.664034] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:28.256 [2024-09-30 14:14:32.664073] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:28.256 [2024-09-30 14:14:32.664127] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:28.256 [2024-09-30 14:14:32.664151] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:28.256 [2024-09-30 14:14:32.664241] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:28.256 [2024-09-30 14:14:32.664253] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:28.256 [2024-09-30 14:14:32.664269] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:15:28.256 [2024-09-30 14:14:32.664300] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:28.256 [2024-09-30 14:14:32.664360] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:15:28.256 [2024-09-30 14:14:32.664368] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:28.256 [2024-09-30 14:14:32.664587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:28.256 [2024-09-30 14:14:32.664745] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:15:28.256 [2024-09-30 14:14:32.664771] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:15:28.256 [2024-09-30 14:14:32.664865] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:28.256 pt1 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.256 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:28.256 "name": "raid_bdev1", 00:15:28.256 "uuid": "022e0569-240f-432d-b1e9-ee8cb9b7e1f2", 00:15:28.256 "strip_size_kb": 0, 00:15:28.256 "state": "online", 00:15:28.256 "raid_level": "raid1", 00:15:28.256 "superblock": true, 00:15:28.256 "num_base_bdevs": 2, 00:15:28.256 "num_base_bdevs_discovered": 1, 00:15:28.256 "num_base_bdevs_operational": 1, 00:15:28.256 "base_bdevs_list": [ 00:15:28.256 { 00:15:28.256 "name": null, 00:15:28.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:28.256 "is_configured": false, 00:15:28.256 "data_offset": 256, 00:15:28.256 "data_size": 7936 00:15:28.256 }, 00:15:28.256 { 00:15:28.257 "name": "pt2", 00:15:28.257 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:28.257 "is_configured": true, 00:15:28.257 "data_offset": 256, 00:15:28.257 "data_size": 7936 00:15:28.257 } 00:15:28.257 ] 00:15:28.257 }' 00:15:28.257 14:14:32 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:28.257 14:14:32 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:28.516 [2024-09-30 14:14:33.117316] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' 022e0569-240f-432d-b1e9-ee8cb9b7e1f2 '!=' 022e0569-240f-432d-b1e9-ee8cb9b7e1f2 ']' 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 97572 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@950 -- # '[' -z 97572 ']' 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # kill -0 97572 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # uname 00:15:28.516 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97572 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:28.776 killing process with pid 97572 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97572' 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@969 -- # kill 97572 00:15:28.776 [2024-09-30 14:14:33.206383] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:28.776 [2024-09-30 14:14:33.206442] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:28.776 [2024-09-30 14:14:33.206471] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:28.776 [2024-09-30 14:14:33.206493] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:15:28.776 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@974 -- # wait 97572 00:15:28.776 [2024-09-30 14:14:33.228374] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:29.036 ************************************ 00:15:29.036 END TEST raid_superblock_test_4k 00:15:29.036 14:14:33 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:15:29.036 00:15:29.036 real 0m4.912s 00:15:29.036 user 0m7.952s 00:15:29.036 sys 0m1.068s 00:15:29.036 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:29.036 14:14:33 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.036 ************************************ 00:15:29.036 14:14:33 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:15:29.036 14:14:33 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:15:29.036 14:14:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:29.036 14:14:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:29.036 14:14:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:29.036 ************************************ 00:15:29.036 START TEST raid_rebuild_test_sb_4k 00:15:29.036 ************************************ 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:29.036 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=97888 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 97888 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 97888 ']' 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:29.037 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:29.037 14:14:33 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.037 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:29.037 Zero copy mechanism will not be used. 00:15:29.037 [2024-09-30 14:14:33.639259] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:29.037 [2024-09-30 14:14:33.639378] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97888 ] 00:15:29.297 [2024-09-30 14:14:33.769156] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:29.297 [2024-09-30 14:14:33.796606] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:29.297 [2024-09-30 14:14:33.842851] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:29.297 [2024-09-30 14:14:33.884990] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:29.297 [2024-09-30 14:14:33.885025] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.867 BaseBdev1_malloc 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.867 [2024-09-30 14:14:34.474812] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:29.867 [2024-09-30 14:14:34.474886] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:29.867 [2024-09-30 14:14:34.474913] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:29.867 [2024-09-30 14:14:34.474936] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:29.867 [2024-09-30 14:14:34.477017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:29.867 [2024-09-30 14:14:34.477062] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:29.867 BaseBdev1 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.867 BaseBdev2_malloc 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.867 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.867 [2024-09-30 14:14:34.519330] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:29.867 [2024-09-30 14:14:34.519526] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:29.867 [2024-09-30 14:14:34.519587] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:29.867 [2024-09-30 14:14:34.519625] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:30.127 [2024-09-30 14:14:34.524195] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:30.127 [2024-09-30 14:14:34.524268] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:30.127 BaseBdev2 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.127 spare_malloc 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.127 spare_delay 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.127 [2024-09-30 14:14:34.562438] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:30.127 [2024-09-30 14:14:34.562597] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:30.127 [2024-09-30 14:14:34.562633] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:30.127 [2024-09-30 14:14:34.562662] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:30.127 [2024-09-30 14:14:34.564746] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:30.127 [2024-09-30 14:14:34.564832] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:30.127 spare 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.127 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.128 [2024-09-30 14:14:34.574468] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:30.128 [2024-09-30 14:14:34.576255] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:30.128 [2024-09-30 14:14:34.576398] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:30.128 [2024-09-30 14:14:34.576412] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:30.128 [2024-09-30 14:14:34.576662] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:30.128 [2024-09-30 14:14:34.576807] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:30.128 [2024-09-30 14:14:34.576822] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:30.128 [2024-09-30 14:14:34.576919] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:30.128 "name": "raid_bdev1", 00:15:30.128 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:30.128 "strip_size_kb": 0, 00:15:30.128 "state": "online", 00:15:30.128 "raid_level": "raid1", 00:15:30.128 "superblock": true, 00:15:30.128 "num_base_bdevs": 2, 00:15:30.128 "num_base_bdevs_discovered": 2, 00:15:30.128 "num_base_bdevs_operational": 2, 00:15:30.128 "base_bdevs_list": [ 00:15:30.128 { 00:15:30.128 "name": "BaseBdev1", 00:15:30.128 "uuid": "cf2368a1-f86d-53a4-b4f8-60ee2279ab7d", 00:15:30.128 "is_configured": true, 00:15:30.128 "data_offset": 256, 00:15:30.128 "data_size": 7936 00:15:30.128 }, 00:15:30.128 { 00:15:30.128 "name": "BaseBdev2", 00:15:30.128 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:30.128 "is_configured": true, 00:15:30.128 "data_offset": 256, 00:15:30.128 "data_size": 7936 00:15:30.128 } 00:15:30.128 ] 00:15:30.128 }' 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:30.128 14:14:34 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.387 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:30.387 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:30.387 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.387 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.647 [2024-09-30 14:14:35.045862] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:30.647 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:30.648 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:30.907 [2024-09-30 14:14:35.345204] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:30.907 /dev/nbd0 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:30.907 1+0 records in 00:15:30.907 1+0 records out 00:15:30.907 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000653131 s, 6.3 MB/s 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:30.907 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:30.908 14:14:35 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:31.477 7936+0 records in 00:15:31.477 7936+0 records out 00:15:31.477 32505856 bytes (33 MB, 31 MiB) copied, 0.594785 s, 54.7 MB/s 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:31.477 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:31.738 [2024-09-30 14:14:36.227533] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.738 [2024-09-30 14:14:36.239638] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.738 "name": "raid_bdev1", 00:15:31.738 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:31.738 "strip_size_kb": 0, 00:15:31.738 "state": "online", 00:15:31.738 "raid_level": "raid1", 00:15:31.738 "superblock": true, 00:15:31.738 "num_base_bdevs": 2, 00:15:31.738 "num_base_bdevs_discovered": 1, 00:15:31.738 "num_base_bdevs_operational": 1, 00:15:31.738 "base_bdevs_list": [ 00:15:31.738 { 00:15:31.738 "name": null, 00:15:31.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.738 "is_configured": false, 00:15:31.738 "data_offset": 0, 00:15:31.738 "data_size": 7936 00:15:31.738 }, 00:15:31.738 { 00:15:31.738 "name": "BaseBdev2", 00:15:31.738 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:31.738 "is_configured": true, 00:15:31.738 "data_offset": 256, 00:15:31.738 "data_size": 7936 00:15:31.738 } 00:15:31.738 ] 00:15:31.738 }' 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.738 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.307 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:32.307 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.308 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.308 [2024-09-30 14:14:36.674934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:32.308 [2024-09-30 14:14:36.679080] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:15:32.308 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.308 14:14:36 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:32.308 [2024-09-30 14:14:36.680955] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.246 "name": "raid_bdev1", 00:15:33.246 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:33.246 "strip_size_kb": 0, 00:15:33.246 "state": "online", 00:15:33.246 "raid_level": "raid1", 00:15:33.246 "superblock": true, 00:15:33.246 "num_base_bdevs": 2, 00:15:33.246 "num_base_bdevs_discovered": 2, 00:15:33.246 "num_base_bdevs_operational": 2, 00:15:33.246 "process": { 00:15:33.246 "type": "rebuild", 00:15:33.246 "target": "spare", 00:15:33.246 "progress": { 00:15:33.246 "blocks": 2560, 00:15:33.246 "percent": 32 00:15:33.246 } 00:15:33.246 }, 00:15:33.246 "base_bdevs_list": [ 00:15:33.246 { 00:15:33.246 "name": "spare", 00:15:33.246 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:33.246 "is_configured": true, 00:15:33.246 "data_offset": 256, 00:15:33.246 "data_size": 7936 00:15:33.246 }, 00:15:33.246 { 00:15:33.246 "name": "BaseBdev2", 00:15:33.246 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:33.246 "is_configured": true, 00:15:33.246 "data_offset": 256, 00:15:33.246 "data_size": 7936 00:15:33.246 } 00:15:33.246 ] 00:15:33.246 }' 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.246 [2024-09-30 14:14:37.845468] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:33.246 [2024-09-30 14:14:37.885335] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:33.246 [2024-09-30 14:14:37.885417] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:33.246 [2024-09-30 14:14:37.885431] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:33.246 [2024-09-30 14:14:37.885443] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.246 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.505 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.505 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.505 "name": "raid_bdev1", 00:15:33.505 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:33.505 "strip_size_kb": 0, 00:15:33.505 "state": "online", 00:15:33.505 "raid_level": "raid1", 00:15:33.505 "superblock": true, 00:15:33.505 "num_base_bdevs": 2, 00:15:33.505 "num_base_bdevs_discovered": 1, 00:15:33.505 "num_base_bdevs_operational": 1, 00:15:33.505 "base_bdevs_list": [ 00:15:33.505 { 00:15:33.505 "name": null, 00:15:33.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.505 "is_configured": false, 00:15:33.505 "data_offset": 0, 00:15:33.505 "data_size": 7936 00:15:33.505 }, 00:15:33.505 { 00:15:33.505 "name": "BaseBdev2", 00:15:33.505 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:33.505 "is_configured": true, 00:15:33.505 "data_offset": 256, 00:15:33.505 "data_size": 7936 00:15:33.505 } 00:15:33.505 ] 00:15:33.505 }' 00:15:33.505 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.505 14:14:37 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.766 "name": "raid_bdev1", 00:15:33.766 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:33.766 "strip_size_kb": 0, 00:15:33.766 "state": "online", 00:15:33.766 "raid_level": "raid1", 00:15:33.766 "superblock": true, 00:15:33.766 "num_base_bdevs": 2, 00:15:33.766 "num_base_bdevs_discovered": 1, 00:15:33.766 "num_base_bdevs_operational": 1, 00:15:33.766 "base_bdevs_list": [ 00:15:33.766 { 00:15:33.766 "name": null, 00:15:33.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.766 "is_configured": false, 00:15:33.766 "data_offset": 0, 00:15:33.766 "data_size": 7936 00:15:33.766 }, 00:15:33.766 { 00:15:33.766 "name": "BaseBdev2", 00:15:33.766 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:33.766 "is_configured": true, 00:15:33.766 "data_offset": 256, 00:15:33.766 "data_size": 7936 00:15:33.766 } 00:15:33.766 ] 00:15:33.766 }' 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:33.766 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.027 [2024-09-30 14:14:38.440702] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:34.027 [2024-09-30 14:14:38.444002] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d400 00:15:34.027 [2024-09-30 14:14:38.445851] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.027 14:14:38 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.965 "name": "raid_bdev1", 00:15:34.965 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:34.965 "strip_size_kb": 0, 00:15:34.965 "state": "online", 00:15:34.965 "raid_level": "raid1", 00:15:34.965 "superblock": true, 00:15:34.965 "num_base_bdevs": 2, 00:15:34.965 "num_base_bdevs_discovered": 2, 00:15:34.965 "num_base_bdevs_operational": 2, 00:15:34.965 "process": { 00:15:34.965 "type": "rebuild", 00:15:34.965 "target": "spare", 00:15:34.965 "progress": { 00:15:34.965 "blocks": 2560, 00:15:34.965 "percent": 32 00:15:34.965 } 00:15:34.965 }, 00:15:34.965 "base_bdevs_list": [ 00:15:34.965 { 00:15:34.965 "name": "spare", 00:15:34.965 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:34.965 "is_configured": true, 00:15:34.965 "data_offset": 256, 00:15:34.965 "data_size": 7936 00:15:34.965 }, 00:15:34.965 { 00:15:34.965 "name": "BaseBdev2", 00:15:34.965 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:34.965 "is_configured": true, 00:15:34.965 "data_offset": 256, 00:15:34.965 "data_size": 7936 00:15:34.965 } 00:15:34.965 ] 00:15:34.965 }' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:34.965 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=560 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.965 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.224 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:35.224 "name": "raid_bdev1", 00:15:35.224 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:35.224 "strip_size_kb": 0, 00:15:35.224 "state": "online", 00:15:35.224 "raid_level": "raid1", 00:15:35.224 "superblock": true, 00:15:35.224 "num_base_bdevs": 2, 00:15:35.224 "num_base_bdevs_discovered": 2, 00:15:35.224 "num_base_bdevs_operational": 2, 00:15:35.224 "process": { 00:15:35.224 "type": "rebuild", 00:15:35.224 "target": "spare", 00:15:35.224 "progress": { 00:15:35.224 "blocks": 2816, 00:15:35.224 "percent": 35 00:15:35.224 } 00:15:35.224 }, 00:15:35.224 "base_bdevs_list": [ 00:15:35.224 { 00:15:35.224 "name": "spare", 00:15:35.224 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:35.224 "is_configured": true, 00:15:35.224 "data_offset": 256, 00:15:35.224 "data_size": 7936 00:15:35.224 }, 00:15:35.224 { 00:15:35.224 "name": "BaseBdev2", 00:15:35.224 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:35.224 "is_configured": true, 00:15:35.224 "data_offset": 256, 00:15:35.224 "data_size": 7936 00:15:35.224 } 00:15:35.224 ] 00:15:35.224 }' 00:15:35.224 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.224 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:35.224 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.224 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:35.225 14:14:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.162 "name": "raid_bdev1", 00:15:36.162 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:36.162 "strip_size_kb": 0, 00:15:36.162 "state": "online", 00:15:36.162 "raid_level": "raid1", 00:15:36.162 "superblock": true, 00:15:36.162 "num_base_bdevs": 2, 00:15:36.162 "num_base_bdevs_discovered": 2, 00:15:36.162 "num_base_bdevs_operational": 2, 00:15:36.162 "process": { 00:15:36.162 "type": "rebuild", 00:15:36.162 "target": "spare", 00:15:36.162 "progress": { 00:15:36.162 "blocks": 5632, 00:15:36.162 "percent": 70 00:15:36.162 } 00:15:36.162 }, 00:15:36.162 "base_bdevs_list": [ 00:15:36.162 { 00:15:36.162 "name": "spare", 00:15:36.162 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:36.162 "is_configured": true, 00:15:36.162 "data_offset": 256, 00:15:36.162 "data_size": 7936 00:15:36.162 }, 00:15:36.162 { 00:15:36.162 "name": "BaseBdev2", 00:15:36.162 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:36.162 "is_configured": true, 00:15:36.162 "data_offset": 256, 00:15:36.162 "data_size": 7936 00:15:36.162 } 00:15:36.162 ] 00:15:36.162 }' 00:15:36.162 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.421 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:36.421 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.421 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:36.421 14:14:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:36.990 [2024-09-30 14:14:41.555686] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:36.990 [2024-09-30 14:14:41.555806] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:36.990 [2024-09-30 14:14:41.555935] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.249 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.508 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.508 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:37.508 "name": "raid_bdev1", 00:15:37.508 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:37.508 "strip_size_kb": 0, 00:15:37.508 "state": "online", 00:15:37.508 "raid_level": "raid1", 00:15:37.508 "superblock": true, 00:15:37.508 "num_base_bdevs": 2, 00:15:37.508 "num_base_bdevs_discovered": 2, 00:15:37.508 "num_base_bdevs_operational": 2, 00:15:37.508 "base_bdevs_list": [ 00:15:37.508 { 00:15:37.508 "name": "spare", 00:15:37.508 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:37.508 "is_configured": true, 00:15:37.508 "data_offset": 256, 00:15:37.508 "data_size": 7936 00:15:37.508 }, 00:15:37.508 { 00:15:37.508 "name": "BaseBdev2", 00:15:37.508 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:37.508 "is_configured": true, 00:15:37.508 "data_offset": 256, 00:15:37.508 "data_size": 7936 00:15:37.508 } 00:15:37.508 ] 00:15:37.508 }' 00:15:37.508 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:37.508 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:37.508 14:14:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:37.508 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:37.508 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:37.509 "name": "raid_bdev1", 00:15:37.509 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:37.509 "strip_size_kb": 0, 00:15:37.509 "state": "online", 00:15:37.509 "raid_level": "raid1", 00:15:37.509 "superblock": true, 00:15:37.509 "num_base_bdevs": 2, 00:15:37.509 "num_base_bdevs_discovered": 2, 00:15:37.509 "num_base_bdevs_operational": 2, 00:15:37.509 "base_bdevs_list": [ 00:15:37.509 { 00:15:37.509 "name": "spare", 00:15:37.509 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:37.509 "is_configured": true, 00:15:37.509 "data_offset": 256, 00:15:37.509 "data_size": 7936 00:15:37.509 }, 00:15:37.509 { 00:15:37.509 "name": "BaseBdev2", 00:15:37.509 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:37.509 "is_configured": true, 00:15:37.509 "data_offset": 256, 00:15:37.509 "data_size": 7936 00:15:37.509 } 00:15:37.509 ] 00:15:37.509 }' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.509 "name": "raid_bdev1", 00:15:37.509 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:37.509 "strip_size_kb": 0, 00:15:37.509 "state": "online", 00:15:37.509 "raid_level": "raid1", 00:15:37.509 "superblock": true, 00:15:37.509 "num_base_bdevs": 2, 00:15:37.509 "num_base_bdevs_discovered": 2, 00:15:37.509 "num_base_bdevs_operational": 2, 00:15:37.509 "base_bdevs_list": [ 00:15:37.509 { 00:15:37.509 "name": "spare", 00:15:37.509 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:37.509 "is_configured": true, 00:15:37.509 "data_offset": 256, 00:15:37.509 "data_size": 7936 00:15:37.509 }, 00:15:37.509 { 00:15:37.509 "name": "BaseBdev2", 00:15:37.509 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:37.509 "is_configured": true, 00:15:37.509 "data_offset": 256, 00:15:37.509 "data_size": 7936 00:15:37.509 } 00:15:37.509 ] 00:15:37.509 }' 00:15:37.509 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.767 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.026 [2024-09-30 14:14:42.565740] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:38.026 [2024-09-30 14:14:42.565815] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:38.026 [2024-09-30 14:14:42.565896] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:38.026 [2024-09-30 14:14:42.565972] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:38.026 [2024-09-30 14:14:42.566039] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:38.026 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:38.285 /dev/nbd0 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:38.285 1+0 records in 00:15:38.285 1+0 records out 00:15:38.285 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000401919 s, 10.2 MB/s 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:38.285 14:14:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:38.545 /dev/nbd1 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:38.545 1+0 records in 00:15:38.545 1+0 records out 00:15:38.545 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000466486 s, 8.8 MB/s 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:38.545 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:38.805 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.065 [2024-09-30 14:14:43.657045] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:39.065 [2024-09-30 14:14:43.657103] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:39.065 [2024-09-30 14:14:43.657127] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:39.065 [2024-09-30 14:14:43.657136] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:39.065 [2024-09-30 14:14:43.659233] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:39.065 [2024-09-30 14:14:43.659272] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:39.065 [2024-09-30 14:14:43.659350] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:39.065 [2024-09-30 14:14:43.659394] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:39.065 [2024-09-30 14:14:43.659547] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:39.065 spare 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.065 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.326 [2024-09-30 14:14:43.759442] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:15:39.326 [2024-09-30 14:14:43.759472] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:39.326 [2024-09-30 14:14:43.759726] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:15:39.326 [2024-09-30 14:14:43.759872] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:15:39.326 [2024-09-30 14:14:43.759889] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:15:39.326 [2024-09-30 14:14:43.760017] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.326 "name": "raid_bdev1", 00:15:39.326 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:39.326 "strip_size_kb": 0, 00:15:39.326 "state": "online", 00:15:39.326 "raid_level": "raid1", 00:15:39.326 "superblock": true, 00:15:39.326 "num_base_bdevs": 2, 00:15:39.326 "num_base_bdevs_discovered": 2, 00:15:39.326 "num_base_bdevs_operational": 2, 00:15:39.326 "base_bdevs_list": [ 00:15:39.326 { 00:15:39.326 "name": "spare", 00:15:39.326 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:39.326 "is_configured": true, 00:15:39.326 "data_offset": 256, 00:15:39.326 "data_size": 7936 00:15:39.326 }, 00:15:39.326 { 00:15:39.326 "name": "BaseBdev2", 00:15:39.326 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:39.326 "is_configured": true, 00:15:39.326 "data_offset": 256, 00:15:39.326 "data_size": 7936 00:15:39.326 } 00:15:39.326 ] 00:15:39.326 }' 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.326 14:14:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:39.587 "name": "raid_bdev1", 00:15:39.587 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:39.587 "strip_size_kb": 0, 00:15:39.587 "state": "online", 00:15:39.587 "raid_level": "raid1", 00:15:39.587 "superblock": true, 00:15:39.587 "num_base_bdevs": 2, 00:15:39.587 "num_base_bdevs_discovered": 2, 00:15:39.587 "num_base_bdevs_operational": 2, 00:15:39.587 "base_bdevs_list": [ 00:15:39.587 { 00:15:39.587 "name": "spare", 00:15:39.587 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:39.587 "is_configured": true, 00:15:39.587 "data_offset": 256, 00:15:39.587 "data_size": 7936 00:15:39.587 }, 00:15:39.587 { 00:15:39.587 "name": "BaseBdev2", 00:15:39.587 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:39.587 "is_configured": true, 00:15:39.587 "data_offset": 256, 00:15:39.587 "data_size": 7936 00:15:39.587 } 00:15:39.587 ] 00:15:39.587 }' 00:15:39.587 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.846 [2024-09-30 14:14:44.371852] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.846 "name": "raid_bdev1", 00:15:39.846 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:39.846 "strip_size_kb": 0, 00:15:39.846 "state": "online", 00:15:39.846 "raid_level": "raid1", 00:15:39.846 "superblock": true, 00:15:39.846 "num_base_bdevs": 2, 00:15:39.846 "num_base_bdevs_discovered": 1, 00:15:39.846 "num_base_bdevs_operational": 1, 00:15:39.846 "base_bdevs_list": [ 00:15:39.846 { 00:15:39.846 "name": null, 00:15:39.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.846 "is_configured": false, 00:15:39.846 "data_offset": 0, 00:15:39.846 "data_size": 7936 00:15:39.846 }, 00:15:39.846 { 00:15:39.846 "name": "BaseBdev2", 00:15:39.846 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:39.846 "is_configured": true, 00:15:39.846 "data_offset": 256, 00:15:39.846 "data_size": 7936 00:15:39.846 } 00:15:39.846 ] 00:15:39.846 }' 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.846 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.416 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:40.416 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:40.416 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.416 [2024-09-30 14:14:44.815518] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:40.416 [2024-09-30 14:14:44.815697] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:40.416 [2024-09-30 14:14:44.815765] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:40.416 [2024-09-30 14:14:44.815833] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:40.416 [2024-09-30 14:14:44.819863] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:15:40.416 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:40.416 14:14:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:40.416 [2024-09-30 14:14:44.821708] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:41.356 "name": "raid_bdev1", 00:15:41.356 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:41.356 "strip_size_kb": 0, 00:15:41.356 "state": "online", 00:15:41.356 "raid_level": "raid1", 00:15:41.356 "superblock": true, 00:15:41.356 "num_base_bdevs": 2, 00:15:41.356 "num_base_bdevs_discovered": 2, 00:15:41.356 "num_base_bdevs_operational": 2, 00:15:41.356 "process": { 00:15:41.356 "type": "rebuild", 00:15:41.356 "target": "spare", 00:15:41.356 "progress": { 00:15:41.356 "blocks": 2560, 00:15:41.356 "percent": 32 00:15:41.356 } 00:15:41.356 }, 00:15:41.356 "base_bdevs_list": [ 00:15:41.356 { 00:15:41.356 "name": "spare", 00:15:41.356 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:41.356 "is_configured": true, 00:15:41.356 "data_offset": 256, 00:15:41.356 "data_size": 7936 00:15:41.356 }, 00:15:41.356 { 00:15:41.356 "name": "BaseBdev2", 00:15:41.356 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:41.356 "is_configured": true, 00:15:41.356 "data_offset": 256, 00:15:41.356 "data_size": 7936 00:15:41.356 } 00:15:41.356 ] 00:15:41.356 }' 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.356 14:14:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.356 [2024-09-30 14:14:45.927188] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:41.616 [2024-09-30 14:14:46.025605] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:41.616 [2024-09-30 14:14:46.025658] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:41.616 [2024-09-30 14:14:46.025672] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:41.616 [2024-09-30 14:14:46.025680] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.616 "name": "raid_bdev1", 00:15:41.616 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:41.616 "strip_size_kb": 0, 00:15:41.616 "state": "online", 00:15:41.616 "raid_level": "raid1", 00:15:41.616 "superblock": true, 00:15:41.616 "num_base_bdevs": 2, 00:15:41.616 "num_base_bdevs_discovered": 1, 00:15:41.616 "num_base_bdevs_operational": 1, 00:15:41.616 "base_bdevs_list": [ 00:15:41.616 { 00:15:41.616 "name": null, 00:15:41.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.616 "is_configured": false, 00:15:41.616 "data_offset": 0, 00:15:41.616 "data_size": 7936 00:15:41.616 }, 00:15:41.616 { 00:15:41.616 "name": "BaseBdev2", 00:15:41.616 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:41.616 "is_configured": true, 00:15:41.616 "data_offset": 256, 00:15:41.616 "data_size": 7936 00:15:41.616 } 00:15:41.616 ] 00:15:41.616 }' 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.616 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.877 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:41.877 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.877 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.877 [2024-09-30 14:14:46.432565] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:41.877 [2024-09-30 14:14:46.432679] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:41.877 [2024-09-30 14:14:46.432717] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:15:41.877 [2024-09-30 14:14:46.432746] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:41.877 [2024-09-30 14:14:46.433162] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:41.877 [2024-09-30 14:14:46.433220] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:41.877 [2024-09-30 14:14:46.433315] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:41.877 [2024-09-30 14:14:46.433360] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:41.877 [2024-09-30 14:14:46.433402] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:41.877 [2024-09-30 14:14:46.433464] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:41.877 [2024-09-30 14:14:46.436551] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1dc0 00:15:41.877 spare 00:15:41.877 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.877 14:14:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:41.877 [2024-09-30 14:14:46.438344] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.815 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.075 "name": "raid_bdev1", 00:15:43.075 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:43.075 "strip_size_kb": 0, 00:15:43.075 "state": "online", 00:15:43.075 "raid_level": "raid1", 00:15:43.075 "superblock": true, 00:15:43.075 "num_base_bdevs": 2, 00:15:43.075 "num_base_bdevs_discovered": 2, 00:15:43.075 "num_base_bdevs_operational": 2, 00:15:43.075 "process": { 00:15:43.075 "type": "rebuild", 00:15:43.075 "target": "spare", 00:15:43.075 "progress": { 00:15:43.075 "blocks": 2560, 00:15:43.075 "percent": 32 00:15:43.075 } 00:15:43.075 }, 00:15:43.075 "base_bdevs_list": [ 00:15:43.075 { 00:15:43.075 "name": "spare", 00:15:43.075 "uuid": "8e8ca7d4-6370-5ae0-a551-b6d3fc73560d", 00:15:43.075 "is_configured": true, 00:15:43.075 "data_offset": 256, 00:15:43.075 "data_size": 7936 00:15:43.075 }, 00:15:43.075 { 00:15:43.075 "name": "BaseBdev2", 00:15:43.075 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:43.075 "is_configured": true, 00:15:43.075 "data_offset": 256, 00:15:43.075 "data_size": 7936 00:15:43.075 } 00:15:43.075 ] 00:15:43.075 }' 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.075 [2024-09-30 14:14:47.595513] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:43.075 [2024-09-30 14:14:47.642184] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:43.075 [2024-09-30 14:14:47.642277] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:43.075 [2024-09-30 14:14:47.642311] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:43.075 [2024-09-30 14:14:47.642330] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.075 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.075 "name": "raid_bdev1", 00:15:43.075 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:43.075 "strip_size_kb": 0, 00:15:43.075 "state": "online", 00:15:43.076 "raid_level": "raid1", 00:15:43.076 "superblock": true, 00:15:43.076 "num_base_bdevs": 2, 00:15:43.076 "num_base_bdevs_discovered": 1, 00:15:43.076 "num_base_bdevs_operational": 1, 00:15:43.076 "base_bdevs_list": [ 00:15:43.076 { 00:15:43.076 "name": null, 00:15:43.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.076 "is_configured": false, 00:15:43.076 "data_offset": 0, 00:15:43.076 "data_size": 7936 00:15:43.076 }, 00:15:43.076 { 00:15:43.076 "name": "BaseBdev2", 00:15:43.076 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:43.076 "is_configured": true, 00:15:43.076 "data_offset": 256, 00:15:43.076 "data_size": 7936 00:15:43.076 } 00:15:43.076 ] 00:15:43.076 }' 00:15:43.076 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.076 14:14:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.645 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.645 "name": "raid_bdev1", 00:15:43.645 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:43.645 "strip_size_kb": 0, 00:15:43.645 "state": "online", 00:15:43.645 "raid_level": "raid1", 00:15:43.645 "superblock": true, 00:15:43.645 "num_base_bdevs": 2, 00:15:43.645 "num_base_bdevs_discovered": 1, 00:15:43.645 "num_base_bdevs_operational": 1, 00:15:43.646 "base_bdevs_list": [ 00:15:43.646 { 00:15:43.646 "name": null, 00:15:43.646 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.646 "is_configured": false, 00:15:43.646 "data_offset": 0, 00:15:43.646 "data_size": 7936 00:15:43.646 }, 00:15:43.646 { 00:15:43.646 "name": "BaseBdev2", 00:15:43.646 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:43.646 "is_configured": true, 00:15:43.646 "data_offset": 256, 00:15:43.646 "data_size": 7936 00:15:43.646 } 00:15:43.646 ] 00:15:43.646 }' 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.646 [2024-09-30 14:14:48.221077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:43.646 [2024-09-30 14:14:48.221170] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:43.646 [2024-09-30 14:14:48.221196] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:15:43.646 [2024-09-30 14:14:48.221205] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:43.646 [2024-09-30 14:14:48.221606] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:43.646 [2024-09-30 14:14:48.221624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:43.646 [2024-09-30 14:14:48.221691] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:43.646 [2024-09-30 14:14:48.221710] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:43.646 [2024-09-30 14:14:48.221722] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:43.646 [2024-09-30 14:14:48.221734] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:43.646 BaseBdev1 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.646 14:14:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:44.584 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.842 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:44.842 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.842 "name": "raid_bdev1", 00:15:44.842 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:44.842 "strip_size_kb": 0, 00:15:44.842 "state": "online", 00:15:44.842 "raid_level": "raid1", 00:15:44.842 "superblock": true, 00:15:44.842 "num_base_bdevs": 2, 00:15:44.842 "num_base_bdevs_discovered": 1, 00:15:44.842 "num_base_bdevs_operational": 1, 00:15:44.842 "base_bdevs_list": [ 00:15:44.842 { 00:15:44.842 "name": null, 00:15:44.842 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.842 "is_configured": false, 00:15:44.842 "data_offset": 0, 00:15:44.842 "data_size": 7936 00:15:44.842 }, 00:15:44.842 { 00:15:44.842 "name": "BaseBdev2", 00:15:44.842 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:44.842 "is_configured": true, 00:15:44.842 "data_offset": 256, 00:15:44.842 "data_size": 7936 00:15:44.842 } 00:15:44.842 ] 00:15:44.843 }' 00:15:44.843 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.843 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:45.102 "name": "raid_bdev1", 00:15:45.102 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:45.102 "strip_size_kb": 0, 00:15:45.102 "state": "online", 00:15:45.102 "raid_level": "raid1", 00:15:45.102 "superblock": true, 00:15:45.102 "num_base_bdevs": 2, 00:15:45.102 "num_base_bdevs_discovered": 1, 00:15:45.102 "num_base_bdevs_operational": 1, 00:15:45.102 "base_bdevs_list": [ 00:15:45.102 { 00:15:45.102 "name": null, 00:15:45.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.102 "is_configured": false, 00:15:45.102 "data_offset": 0, 00:15:45.102 "data_size": 7936 00:15:45.102 }, 00:15:45.102 { 00:15:45.102 "name": "BaseBdev2", 00:15:45.102 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:45.102 "is_configured": true, 00:15:45.102 "data_offset": 256, 00:15:45.102 "data_size": 7936 00:15:45.102 } 00:15:45.102 ] 00:15:45.102 }' 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:45.102 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # local es=0 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.362 [2024-09-30 14:14:49.778685] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:45.362 [2024-09-30 14:14:49.778788] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:45.362 [2024-09-30 14:14:49.778802] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:45.362 request: 00:15:45.362 { 00:15:45.362 "base_bdev": "BaseBdev1", 00:15:45.362 "raid_bdev": "raid_bdev1", 00:15:45.362 "method": "bdev_raid_add_base_bdev", 00:15:45.362 "req_id": 1 00:15:45.362 } 00:15:45.362 Got JSON-RPC error response 00:15:45.362 response: 00:15:45.362 { 00:15:45.362 "code": -22, 00:15:45.362 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:45.362 } 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # es=1 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:45.362 14:14:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.300 "name": "raid_bdev1", 00:15:46.300 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:46.300 "strip_size_kb": 0, 00:15:46.300 "state": "online", 00:15:46.300 "raid_level": "raid1", 00:15:46.300 "superblock": true, 00:15:46.300 "num_base_bdevs": 2, 00:15:46.300 "num_base_bdevs_discovered": 1, 00:15:46.300 "num_base_bdevs_operational": 1, 00:15:46.300 "base_bdevs_list": [ 00:15:46.300 { 00:15:46.300 "name": null, 00:15:46.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.300 "is_configured": false, 00:15:46.300 "data_offset": 0, 00:15:46.300 "data_size": 7936 00:15:46.300 }, 00:15:46.300 { 00:15:46.300 "name": "BaseBdev2", 00:15:46.300 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:46.300 "is_configured": true, 00:15:46.300 "data_offset": 256, 00:15:46.300 "data_size": 7936 00:15:46.300 } 00:15:46.300 ] 00:15:46.300 }' 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.300 14:14:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.870 "name": "raid_bdev1", 00:15:46.870 "uuid": "a5f25353-b02c-461e-b9b8-1a239858e374", 00:15:46.870 "strip_size_kb": 0, 00:15:46.870 "state": "online", 00:15:46.870 "raid_level": "raid1", 00:15:46.870 "superblock": true, 00:15:46.870 "num_base_bdevs": 2, 00:15:46.870 "num_base_bdevs_discovered": 1, 00:15:46.870 "num_base_bdevs_operational": 1, 00:15:46.870 "base_bdevs_list": [ 00:15:46.870 { 00:15:46.870 "name": null, 00:15:46.870 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.870 "is_configured": false, 00:15:46.870 "data_offset": 0, 00:15:46.870 "data_size": 7936 00:15:46.870 }, 00:15:46.870 { 00:15:46.870 "name": "BaseBdev2", 00:15:46.870 "uuid": "815e4d9c-3b60-5bb1-a9e5-11a1295b9b63", 00:15:46.870 "is_configured": true, 00:15:46.870 "data_offset": 256, 00:15:46.870 "data_size": 7936 00:15:46.870 } 00:15:46.870 ] 00:15:46.870 }' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 97888 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 97888 ']' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 97888 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97888 00:15:46.870 killing process with pid 97888 00:15:46.870 Received shutdown signal, test time was about 60.000000 seconds 00:15:46.870 00:15:46.870 Latency(us) 00:15:46.870 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:46.870 =================================================================================================================== 00:15:46.870 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97888' 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@969 -- # kill 97888 00:15:46.870 [2024-09-30 14:14:51.406349] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:46.870 [2024-09-30 14:14:51.406447] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:46.870 [2024-09-30 14:14:51.406501] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:46.870 [2024-09-30 14:14:51.406513] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:15:46.870 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@974 -- # wait 97888 00:15:46.870 [2024-09-30 14:14:51.437777] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:47.131 ************************************ 00:15:47.131 END TEST raid_rebuild_test_sb_4k 00:15:47.131 ************************************ 00:15:47.131 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:15:47.131 00:15:47.131 real 0m18.121s 00:15:47.131 user 0m23.901s 00:15:47.131 sys 0m2.608s 00:15:47.131 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:47.131 14:14:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.131 14:14:51 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:15:47.131 14:14:51 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:15:47.131 14:14:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:15:47.131 14:14:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:47.131 14:14:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:47.131 ************************************ 00:15:47.131 START TEST raid_state_function_test_sb_md_separate 00:15:47.131 ************************************ 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=98564 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98564' 00:15:47.131 Process raid pid: 98564 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 98564 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 98564 ']' 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:47.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:47.131 14:14:51 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.392 [2024-09-30 14:14:51.845879] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:47.392 [2024-09-30 14:14:51.846089] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:47.392 [2024-09-30 14:14:51.979813] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:47.392 [2024-09-30 14:14:52.010255] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:47.652 [2024-09-30 14:14:52.058710] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:47.652 [2024-09-30 14:14:52.100868] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.652 [2024-09-30 14:14:52.100978] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.221 [2024-09-30 14:14:52.686340] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:48.221 [2024-09-30 14:14:52.686395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:48.221 [2024-09-30 14:14:52.686407] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:48.221 [2024-09-30 14:14:52.686414] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.221 "name": "Existed_Raid", 00:15:48.221 "uuid": "5ea8d462-cd19-405a-981a-72bbb69037f2", 00:15:48.221 "strip_size_kb": 0, 00:15:48.221 "state": "configuring", 00:15:48.221 "raid_level": "raid1", 00:15:48.221 "superblock": true, 00:15:48.221 "num_base_bdevs": 2, 00:15:48.221 "num_base_bdevs_discovered": 0, 00:15:48.221 "num_base_bdevs_operational": 2, 00:15:48.221 "base_bdevs_list": [ 00:15:48.221 { 00:15:48.221 "name": "BaseBdev1", 00:15:48.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.221 "is_configured": false, 00:15:48.221 "data_offset": 0, 00:15:48.221 "data_size": 0 00:15:48.221 }, 00:15:48.221 { 00:15:48.221 "name": "BaseBdev2", 00:15:48.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.221 "is_configured": false, 00:15:48.221 "data_offset": 0, 00:15:48.221 "data_size": 0 00:15:48.221 } 00:15:48.221 ] 00:15:48.221 }' 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.221 14:14:52 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.481 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:48.481 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.481 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.741 [2024-09-30 14:14:53.141455] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:48.741 [2024-09-30 14:14:53.141563] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.741 [2024-09-30 14:14:53.153438] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:48.741 [2024-09-30 14:14:53.153522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:48.741 [2024-09-30 14:14:53.153550] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:48.741 [2024-09-30 14:14:53.153568] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.741 [2024-09-30 14:14:53.174657] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:48.741 BaseBdev1 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.741 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.741 [ 00:15:48.741 { 00:15:48.741 "name": "BaseBdev1", 00:15:48.741 "aliases": [ 00:15:48.741 "4447c1c0-c710-478f-9baa-7fd00e399e10" 00:15:48.741 ], 00:15:48.741 "product_name": "Malloc disk", 00:15:48.741 "block_size": 4096, 00:15:48.741 "num_blocks": 8192, 00:15:48.741 "uuid": "4447c1c0-c710-478f-9baa-7fd00e399e10", 00:15:48.741 "md_size": 32, 00:15:48.741 "md_interleave": false, 00:15:48.741 "dif_type": 0, 00:15:48.741 "assigned_rate_limits": { 00:15:48.741 "rw_ios_per_sec": 0, 00:15:48.741 "rw_mbytes_per_sec": 0, 00:15:48.741 "r_mbytes_per_sec": 0, 00:15:48.741 "w_mbytes_per_sec": 0 00:15:48.741 }, 00:15:48.741 "claimed": true, 00:15:48.741 "claim_type": "exclusive_write", 00:15:48.741 "zoned": false, 00:15:48.741 "supported_io_types": { 00:15:48.741 "read": true, 00:15:48.741 "write": true, 00:15:48.741 "unmap": true, 00:15:48.742 "flush": true, 00:15:48.742 "reset": true, 00:15:48.742 "nvme_admin": false, 00:15:48.742 "nvme_io": false, 00:15:48.742 "nvme_io_md": false, 00:15:48.742 "write_zeroes": true, 00:15:48.742 "zcopy": true, 00:15:48.742 "get_zone_info": false, 00:15:48.742 "zone_management": false, 00:15:48.742 "zone_append": false, 00:15:48.742 "compare": false, 00:15:48.742 "compare_and_write": false, 00:15:48.742 "abort": true, 00:15:48.742 "seek_hole": false, 00:15:48.742 "seek_data": false, 00:15:48.742 "copy": true, 00:15:48.742 "nvme_iov_md": false 00:15:48.742 }, 00:15:48.742 "memory_domains": [ 00:15:48.742 { 00:15:48.742 "dma_device_id": "system", 00:15:48.742 "dma_device_type": 1 00:15:48.742 }, 00:15:48.742 { 00:15:48.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.742 "dma_device_type": 2 00:15:48.742 } 00:15:48.742 ], 00:15:48.742 "driver_specific": {} 00:15:48.742 } 00:15:48.742 ] 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.742 "name": "Existed_Raid", 00:15:48.742 "uuid": "33c6a5fe-b537-42b6-be92-43a316181de7", 00:15:48.742 "strip_size_kb": 0, 00:15:48.742 "state": "configuring", 00:15:48.742 "raid_level": "raid1", 00:15:48.742 "superblock": true, 00:15:48.742 "num_base_bdevs": 2, 00:15:48.742 "num_base_bdevs_discovered": 1, 00:15:48.742 "num_base_bdevs_operational": 2, 00:15:48.742 "base_bdevs_list": [ 00:15:48.742 { 00:15:48.742 "name": "BaseBdev1", 00:15:48.742 "uuid": "4447c1c0-c710-478f-9baa-7fd00e399e10", 00:15:48.742 "is_configured": true, 00:15:48.742 "data_offset": 256, 00:15:48.742 "data_size": 7936 00:15:48.742 }, 00:15:48.742 { 00:15:48.742 "name": "BaseBdev2", 00:15:48.742 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.742 "is_configured": false, 00:15:48.742 "data_offset": 0, 00:15:48.742 "data_size": 0 00:15:48.742 } 00:15:48.742 ] 00:15:48.742 }' 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.742 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.002 [2024-09-30 14:14:53.605971] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:49.002 [2024-09-30 14:14:53.606065] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.002 [2024-09-30 14:14:53.618025] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:49.002 [2024-09-30 14:14:53.619827] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:49.002 [2024-09-30 14:14:53.619897] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.002 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.262 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.262 "name": "Existed_Raid", 00:15:49.262 "uuid": "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a", 00:15:49.262 "strip_size_kb": 0, 00:15:49.262 "state": "configuring", 00:15:49.262 "raid_level": "raid1", 00:15:49.262 "superblock": true, 00:15:49.262 "num_base_bdevs": 2, 00:15:49.262 "num_base_bdevs_discovered": 1, 00:15:49.262 "num_base_bdevs_operational": 2, 00:15:49.262 "base_bdevs_list": [ 00:15:49.262 { 00:15:49.262 "name": "BaseBdev1", 00:15:49.262 "uuid": "4447c1c0-c710-478f-9baa-7fd00e399e10", 00:15:49.262 "is_configured": true, 00:15:49.262 "data_offset": 256, 00:15:49.262 "data_size": 7936 00:15:49.262 }, 00:15:49.262 { 00:15:49.262 "name": "BaseBdev2", 00:15:49.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:49.262 "is_configured": false, 00:15:49.262 "data_offset": 0, 00:15:49.262 "data_size": 0 00:15:49.262 } 00:15:49.262 ] 00:15:49.262 }' 00:15:49.262 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.262 14:14:53 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.522 [2024-09-30 14:14:54.076796] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:49.522 [2024-09-30 14:14:54.077469] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:49.522 BaseBdev2 00:15:49.522 [2024-09-30 14:14:54.077704] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:49.522 [2024-09-30 14:14:54.077984] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:49.522 [2024-09-30 14:14:54.078303] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:49.522 [2024-09-30 14:14:54.078355] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:15:49.522 [2024-09-30 14:14:54.078661] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.522 [ 00:15:49.522 { 00:15:49.522 "name": "BaseBdev2", 00:15:49.522 "aliases": [ 00:15:49.522 "a4e54d5e-54a8-43c9-812a-5b554b3bdd4a" 00:15:49.522 ], 00:15:49.522 "product_name": "Malloc disk", 00:15:49.522 "block_size": 4096, 00:15:49.522 "num_blocks": 8192, 00:15:49.522 "uuid": "a4e54d5e-54a8-43c9-812a-5b554b3bdd4a", 00:15:49.522 "md_size": 32, 00:15:49.522 "md_interleave": false, 00:15:49.522 "dif_type": 0, 00:15:49.522 "assigned_rate_limits": { 00:15:49.522 "rw_ios_per_sec": 0, 00:15:49.522 "rw_mbytes_per_sec": 0, 00:15:49.522 "r_mbytes_per_sec": 0, 00:15:49.522 "w_mbytes_per_sec": 0 00:15:49.522 }, 00:15:49.522 "claimed": true, 00:15:49.522 "claim_type": "exclusive_write", 00:15:49.522 "zoned": false, 00:15:49.522 "supported_io_types": { 00:15:49.522 "read": true, 00:15:49.522 "write": true, 00:15:49.522 "unmap": true, 00:15:49.522 "flush": true, 00:15:49.522 "reset": true, 00:15:49.522 "nvme_admin": false, 00:15:49.522 "nvme_io": false, 00:15:49.522 "nvme_io_md": false, 00:15:49.522 "write_zeroes": true, 00:15:49.522 "zcopy": true, 00:15:49.522 "get_zone_info": false, 00:15:49.522 "zone_management": false, 00:15:49.522 "zone_append": false, 00:15:49.522 "compare": false, 00:15:49.522 "compare_and_write": false, 00:15:49.522 "abort": true, 00:15:49.522 "seek_hole": false, 00:15:49.522 "seek_data": false, 00:15:49.522 "copy": true, 00:15:49.522 "nvme_iov_md": false 00:15:49.522 }, 00:15:49.522 "memory_domains": [ 00:15:49.522 { 00:15:49.522 "dma_device_id": "system", 00:15:49.522 "dma_device_type": 1 00:15:49.522 }, 00:15:49.522 { 00:15:49.522 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:49.522 "dma_device_type": 2 00:15:49.522 } 00:15:49.522 ], 00:15:49.522 "driver_specific": {} 00:15:49.522 } 00:15:49.522 ] 00:15:49.522 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.523 "name": "Existed_Raid", 00:15:49.523 "uuid": "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a", 00:15:49.523 "strip_size_kb": 0, 00:15:49.523 "state": "online", 00:15:49.523 "raid_level": "raid1", 00:15:49.523 "superblock": true, 00:15:49.523 "num_base_bdevs": 2, 00:15:49.523 "num_base_bdevs_discovered": 2, 00:15:49.523 "num_base_bdevs_operational": 2, 00:15:49.523 "base_bdevs_list": [ 00:15:49.523 { 00:15:49.523 "name": "BaseBdev1", 00:15:49.523 "uuid": "4447c1c0-c710-478f-9baa-7fd00e399e10", 00:15:49.523 "is_configured": true, 00:15:49.523 "data_offset": 256, 00:15:49.523 "data_size": 7936 00:15:49.523 }, 00:15:49.523 { 00:15:49.523 "name": "BaseBdev2", 00:15:49.523 "uuid": "a4e54d5e-54a8-43c9-812a-5b554b3bdd4a", 00:15:49.523 "is_configured": true, 00:15:49.523 "data_offset": 256, 00:15:49.523 "data_size": 7936 00:15:49.523 } 00:15:49.523 ] 00:15:49.523 }' 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.523 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:50.093 [2024-09-30 14:14:54.572146] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.093 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:50.093 "name": "Existed_Raid", 00:15:50.093 "aliases": [ 00:15:50.093 "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a" 00:15:50.093 ], 00:15:50.093 "product_name": "Raid Volume", 00:15:50.093 "block_size": 4096, 00:15:50.093 "num_blocks": 7936, 00:15:50.093 "uuid": "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a", 00:15:50.093 "md_size": 32, 00:15:50.093 "md_interleave": false, 00:15:50.093 "dif_type": 0, 00:15:50.093 "assigned_rate_limits": { 00:15:50.093 "rw_ios_per_sec": 0, 00:15:50.093 "rw_mbytes_per_sec": 0, 00:15:50.093 "r_mbytes_per_sec": 0, 00:15:50.093 "w_mbytes_per_sec": 0 00:15:50.093 }, 00:15:50.093 "claimed": false, 00:15:50.093 "zoned": false, 00:15:50.093 "supported_io_types": { 00:15:50.093 "read": true, 00:15:50.093 "write": true, 00:15:50.093 "unmap": false, 00:15:50.093 "flush": false, 00:15:50.093 "reset": true, 00:15:50.093 "nvme_admin": false, 00:15:50.093 "nvme_io": false, 00:15:50.093 "nvme_io_md": false, 00:15:50.093 "write_zeroes": true, 00:15:50.093 "zcopy": false, 00:15:50.093 "get_zone_info": false, 00:15:50.093 "zone_management": false, 00:15:50.093 "zone_append": false, 00:15:50.093 "compare": false, 00:15:50.093 "compare_and_write": false, 00:15:50.093 "abort": false, 00:15:50.093 "seek_hole": false, 00:15:50.093 "seek_data": false, 00:15:50.093 "copy": false, 00:15:50.093 "nvme_iov_md": false 00:15:50.093 }, 00:15:50.093 "memory_domains": [ 00:15:50.093 { 00:15:50.093 "dma_device_id": "system", 00:15:50.093 "dma_device_type": 1 00:15:50.093 }, 00:15:50.093 { 00:15:50.093 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.093 "dma_device_type": 2 00:15:50.093 }, 00:15:50.093 { 00:15:50.093 "dma_device_id": "system", 00:15:50.093 "dma_device_type": 1 00:15:50.093 }, 00:15:50.093 { 00:15:50.093 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.094 "dma_device_type": 2 00:15:50.094 } 00:15:50.094 ], 00:15:50.094 "driver_specific": { 00:15:50.094 "raid": { 00:15:50.094 "uuid": "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a", 00:15:50.094 "strip_size_kb": 0, 00:15:50.094 "state": "online", 00:15:50.094 "raid_level": "raid1", 00:15:50.094 "superblock": true, 00:15:50.094 "num_base_bdevs": 2, 00:15:50.094 "num_base_bdevs_discovered": 2, 00:15:50.094 "num_base_bdevs_operational": 2, 00:15:50.094 "base_bdevs_list": [ 00:15:50.094 { 00:15:50.094 "name": "BaseBdev1", 00:15:50.094 "uuid": "4447c1c0-c710-478f-9baa-7fd00e399e10", 00:15:50.094 "is_configured": true, 00:15:50.094 "data_offset": 256, 00:15:50.094 "data_size": 7936 00:15:50.094 }, 00:15:50.094 { 00:15:50.094 "name": "BaseBdev2", 00:15:50.094 "uuid": "a4e54d5e-54a8-43c9-812a-5b554b3bdd4a", 00:15:50.094 "is_configured": true, 00:15:50.094 "data_offset": 256, 00:15:50.094 "data_size": 7936 00:15:50.094 } 00:15:50.094 ] 00:15:50.094 } 00:15:50.094 } 00:15:50.094 }' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:50.094 BaseBdev2' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.094 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.370 [2024-09-30 14:14:54.799612] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.370 "name": "Existed_Raid", 00:15:50.370 "uuid": "45bc1ced-6c7a-4c6a-95d1-b8f590e66d9a", 00:15:50.370 "strip_size_kb": 0, 00:15:50.370 "state": "online", 00:15:50.370 "raid_level": "raid1", 00:15:50.370 "superblock": true, 00:15:50.370 "num_base_bdevs": 2, 00:15:50.370 "num_base_bdevs_discovered": 1, 00:15:50.370 "num_base_bdevs_operational": 1, 00:15:50.370 "base_bdevs_list": [ 00:15:50.370 { 00:15:50.370 "name": null, 00:15:50.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.370 "is_configured": false, 00:15:50.370 "data_offset": 0, 00:15:50.370 "data_size": 7936 00:15:50.370 }, 00:15:50.370 { 00:15:50.370 "name": "BaseBdev2", 00:15:50.370 "uuid": "a4e54d5e-54a8-43c9-812a-5b554b3bdd4a", 00:15:50.370 "is_configured": true, 00:15:50.370 "data_offset": 256, 00:15:50.370 "data_size": 7936 00:15:50.370 } 00:15:50.370 ] 00:15:50.370 }' 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.370 14:14:54 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.630 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.890 [2024-09-30 14:14:55.294563] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:50.890 [2024-09-30 14:14:55.294705] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:50.890 [2024-09-30 14:14:55.307092] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:50.890 [2024-09-30 14:14:55.307216] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:50.890 [2024-09-30 14:14:55.307261] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 98564 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 98564 ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 98564 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98564 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:50.890 killing process with pid 98564 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98564' 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 98564 00:15:50.890 [2024-09-30 14:14:55.404339] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:50.890 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 98564 00:15:50.890 [2024-09-30 14:14:55.405304] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:51.151 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:15:51.151 00:15:51.151 real 0m3.904s 00:15:51.151 user 0m6.074s 00:15:51.151 sys 0m0.860s 00:15:51.151 ************************************ 00:15:51.151 END TEST raid_state_function_test_sb_md_separate 00:15:51.151 ************************************ 00:15:51.151 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:51.151 14:14:55 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.151 14:14:55 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:15:51.151 14:14:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:51.151 14:14:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:51.151 14:14:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:51.151 ************************************ 00:15:51.151 START TEST raid_superblock_test_md_separate 00:15:51.151 ************************************ 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=98800 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 98800 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:51.151 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@831 -- # '[' -z 98800 ']' 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:51.151 14:14:55 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.420 [2024-09-30 14:14:55.823140] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:51.420 [2024-09-30 14:14:55.823263] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98800 ] 00:15:51.420 [2024-09-30 14:14:55.954770] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:51.420 [2024-09-30 14:14:55.981445] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:51.420 [2024-09-30 14:14:56.026390] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:51.420 [2024-09-30 14:14:56.069343] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:51.698 [2024-09-30 14:14:56.069465] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:52.301 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 malloc1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 [2024-09-30 14:14:56.688757] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:52.302 [2024-09-30 14:14:56.688916] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:52.302 [2024-09-30 14:14:56.688965] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:52.302 [2024-09-30 14:14:56.688999] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:52.302 [2024-09-30 14:14:56.690905] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:52.302 [2024-09-30 14:14:56.690976] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:52.302 pt1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 malloc2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 [2024-09-30 14:14:56.739623] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:52.302 [2024-09-30 14:14:56.739831] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:52.302 [2024-09-30 14:14:56.739921] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:52.302 [2024-09-30 14:14:56.740019] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:52.302 [2024-09-30 14:14:56.743823] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:52.302 [2024-09-30 14:14:56.743922] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:52.302 pt2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 [2024-09-30 14:14:56.752217] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:52.302 [2024-09-30 14:14:56.754503] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:52.302 [2024-09-30 14:14:56.754721] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:52.302 [2024-09-30 14:14:56.754784] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:52.302 [2024-09-30 14:14:56.754901] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:52.302 [2024-09-30 14:14:56.755069] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:52.302 [2024-09-30 14:14:56.755118] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:52.302 [2024-09-30 14:14:56.755258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.302 "name": "raid_bdev1", 00:15:52.302 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:52.302 "strip_size_kb": 0, 00:15:52.302 "state": "online", 00:15:52.302 "raid_level": "raid1", 00:15:52.302 "superblock": true, 00:15:52.302 "num_base_bdevs": 2, 00:15:52.302 "num_base_bdevs_discovered": 2, 00:15:52.302 "num_base_bdevs_operational": 2, 00:15:52.302 "base_bdevs_list": [ 00:15:52.302 { 00:15:52.302 "name": "pt1", 00:15:52.302 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:52.302 "is_configured": true, 00:15:52.302 "data_offset": 256, 00:15:52.302 "data_size": 7936 00:15:52.302 }, 00:15:52.302 { 00:15:52.302 "name": "pt2", 00:15:52.302 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:52.302 "is_configured": true, 00:15:52.302 "data_offset": 256, 00:15:52.302 "data_size": 7936 00:15:52.302 } 00:15:52.302 ] 00:15:52.302 }' 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.302 14:14:56 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:52.580 [2024-09-30 14:14:57.207657] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.580 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:52.840 "name": "raid_bdev1", 00:15:52.840 "aliases": [ 00:15:52.840 "7e212a68-c9ac-4544-865c-daa2eb190c55" 00:15:52.840 ], 00:15:52.840 "product_name": "Raid Volume", 00:15:52.840 "block_size": 4096, 00:15:52.840 "num_blocks": 7936, 00:15:52.840 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:52.840 "md_size": 32, 00:15:52.840 "md_interleave": false, 00:15:52.840 "dif_type": 0, 00:15:52.840 "assigned_rate_limits": { 00:15:52.840 "rw_ios_per_sec": 0, 00:15:52.840 "rw_mbytes_per_sec": 0, 00:15:52.840 "r_mbytes_per_sec": 0, 00:15:52.840 "w_mbytes_per_sec": 0 00:15:52.840 }, 00:15:52.840 "claimed": false, 00:15:52.840 "zoned": false, 00:15:52.840 "supported_io_types": { 00:15:52.840 "read": true, 00:15:52.840 "write": true, 00:15:52.840 "unmap": false, 00:15:52.840 "flush": false, 00:15:52.840 "reset": true, 00:15:52.840 "nvme_admin": false, 00:15:52.840 "nvme_io": false, 00:15:52.840 "nvme_io_md": false, 00:15:52.840 "write_zeroes": true, 00:15:52.840 "zcopy": false, 00:15:52.840 "get_zone_info": false, 00:15:52.840 "zone_management": false, 00:15:52.840 "zone_append": false, 00:15:52.840 "compare": false, 00:15:52.840 "compare_and_write": false, 00:15:52.840 "abort": false, 00:15:52.840 "seek_hole": false, 00:15:52.840 "seek_data": false, 00:15:52.840 "copy": false, 00:15:52.840 "nvme_iov_md": false 00:15:52.840 }, 00:15:52.840 "memory_domains": [ 00:15:52.840 { 00:15:52.840 "dma_device_id": "system", 00:15:52.840 "dma_device_type": 1 00:15:52.840 }, 00:15:52.840 { 00:15:52.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:52.840 "dma_device_type": 2 00:15:52.840 }, 00:15:52.840 { 00:15:52.840 "dma_device_id": "system", 00:15:52.840 "dma_device_type": 1 00:15:52.840 }, 00:15:52.840 { 00:15:52.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:52.840 "dma_device_type": 2 00:15:52.840 } 00:15:52.840 ], 00:15:52.840 "driver_specific": { 00:15:52.840 "raid": { 00:15:52.840 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:52.840 "strip_size_kb": 0, 00:15:52.840 "state": "online", 00:15:52.840 "raid_level": "raid1", 00:15:52.840 "superblock": true, 00:15:52.840 "num_base_bdevs": 2, 00:15:52.840 "num_base_bdevs_discovered": 2, 00:15:52.840 "num_base_bdevs_operational": 2, 00:15:52.840 "base_bdevs_list": [ 00:15:52.840 { 00:15:52.840 "name": "pt1", 00:15:52.840 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:52.840 "is_configured": true, 00:15:52.840 "data_offset": 256, 00:15:52.840 "data_size": 7936 00:15:52.840 }, 00:15:52.840 { 00:15:52.840 "name": "pt2", 00:15:52.840 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:52.840 "is_configured": true, 00:15:52.840 "data_offset": 256, 00:15:52.840 "data_size": 7936 00:15:52.840 } 00:15:52.840 ] 00:15:52.840 } 00:15:52.840 } 00:15:52.840 }' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:52.840 pt2' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.840 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:52.841 [2024-09-30 14:14:57.427170] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7e212a68-c9ac-4544-865c-daa2eb190c55 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 7e212a68-c9ac-4544-865c-daa2eb190c55 ']' 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.841 [2024-09-30 14:14:57.478857] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:52.841 [2024-09-30 14:14:57.478879] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.841 [2024-09-30 14:14:57.478946] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.841 [2024-09-30 14:14:57.478996] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:52.841 [2024-09-30 14:14:57.479008] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.841 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.100 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 [2024-09-30 14:14:57.618616] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:53.101 [2024-09-30 14:14:57.620469] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:53.101 [2024-09-30 14:14:57.620577] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:53.101 [2024-09-30 14:14:57.620657] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:53.101 [2024-09-30 14:14:57.620693] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:53.101 [2024-09-30 14:14:57.620722] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:15:53.101 request: 00:15:53.101 { 00:15:53.101 "name": "raid_bdev1", 00:15:53.101 "raid_level": "raid1", 00:15:53.101 "base_bdevs": [ 00:15:53.101 "malloc1", 00:15:53.101 "malloc2" 00:15:53.101 ], 00:15:53.101 "superblock": false, 00:15:53.101 "method": "bdev_raid_create", 00:15:53.101 "req_id": 1 00:15:53.101 } 00:15:53.101 Got JSON-RPC error response 00:15:53.101 response: 00:15:53.101 { 00:15:53.101 "code": -17, 00:15:53.101 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:53.101 } 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # es=1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 [2024-09-30 14:14:57.686491] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:53.101 [2024-09-30 14:14:57.686581] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.101 [2024-09-30 14:14:57.686608] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:53.101 [2024-09-30 14:14:57.686637] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.101 [2024-09-30 14:14:57.688529] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.101 [2024-09-30 14:14:57.688597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:53.101 [2024-09-30 14:14:57.688655] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:53.101 [2024-09-30 14:14:57.688692] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:53.101 pt1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.101 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.101 "name": "raid_bdev1", 00:15:53.101 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:53.101 "strip_size_kb": 0, 00:15:53.101 "state": "configuring", 00:15:53.101 "raid_level": "raid1", 00:15:53.101 "superblock": true, 00:15:53.101 "num_base_bdevs": 2, 00:15:53.101 "num_base_bdevs_discovered": 1, 00:15:53.101 "num_base_bdevs_operational": 2, 00:15:53.101 "base_bdevs_list": [ 00:15:53.101 { 00:15:53.101 "name": "pt1", 00:15:53.101 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:53.101 "is_configured": true, 00:15:53.101 "data_offset": 256, 00:15:53.101 "data_size": 7936 00:15:53.101 }, 00:15:53.101 { 00:15:53.101 "name": null, 00:15:53.101 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:53.101 "is_configured": false, 00:15:53.101 "data_offset": 256, 00:15:53.102 "data_size": 7936 00:15:53.102 } 00:15:53.102 ] 00:15:53.102 }' 00:15:53.102 14:14:57 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.102 14:14:57 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.671 [2024-09-30 14:14:58.153659] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:53.671 [2024-09-30 14:14:58.153752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.671 [2024-09-30 14:14:58.153785] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:53.671 [2024-09-30 14:14:58.153809] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.671 [2024-09-30 14:14:58.153961] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.671 [2024-09-30 14:14:58.154014] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:53.671 [2024-09-30 14:14:58.154080] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:53.671 [2024-09-30 14:14:58.154127] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:53.671 [2024-09-30 14:14:58.154220] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:53.671 [2024-09-30 14:14:58.154256] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:53.671 [2024-09-30 14:14:58.154333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:53.671 [2024-09-30 14:14:58.154443] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:53.671 [2024-09-30 14:14:58.154475] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:15:53.671 [2024-09-30 14:14:58.154583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.671 pt2 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.671 "name": "raid_bdev1", 00:15:53.671 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:53.671 "strip_size_kb": 0, 00:15:53.671 "state": "online", 00:15:53.671 "raid_level": "raid1", 00:15:53.671 "superblock": true, 00:15:53.671 "num_base_bdevs": 2, 00:15:53.671 "num_base_bdevs_discovered": 2, 00:15:53.671 "num_base_bdevs_operational": 2, 00:15:53.671 "base_bdevs_list": [ 00:15:53.671 { 00:15:53.671 "name": "pt1", 00:15:53.671 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:53.671 "is_configured": true, 00:15:53.671 "data_offset": 256, 00:15:53.671 "data_size": 7936 00:15:53.671 }, 00:15:53.671 { 00:15:53.671 "name": "pt2", 00:15:53.671 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:53.671 "is_configured": true, 00:15:53.671 "data_offset": 256, 00:15:53.671 "data_size": 7936 00:15:53.671 } 00:15:53.671 ] 00:15:53.671 }' 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.671 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.931 [2024-09-30 14:14:58.525218] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:53.931 "name": "raid_bdev1", 00:15:53.931 "aliases": [ 00:15:53.931 "7e212a68-c9ac-4544-865c-daa2eb190c55" 00:15:53.931 ], 00:15:53.931 "product_name": "Raid Volume", 00:15:53.931 "block_size": 4096, 00:15:53.931 "num_blocks": 7936, 00:15:53.931 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:53.931 "md_size": 32, 00:15:53.931 "md_interleave": false, 00:15:53.931 "dif_type": 0, 00:15:53.931 "assigned_rate_limits": { 00:15:53.931 "rw_ios_per_sec": 0, 00:15:53.931 "rw_mbytes_per_sec": 0, 00:15:53.931 "r_mbytes_per_sec": 0, 00:15:53.931 "w_mbytes_per_sec": 0 00:15:53.931 }, 00:15:53.931 "claimed": false, 00:15:53.931 "zoned": false, 00:15:53.931 "supported_io_types": { 00:15:53.931 "read": true, 00:15:53.931 "write": true, 00:15:53.931 "unmap": false, 00:15:53.931 "flush": false, 00:15:53.931 "reset": true, 00:15:53.931 "nvme_admin": false, 00:15:53.931 "nvme_io": false, 00:15:53.931 "nvme_io_md": false, 00:15:53.931 "write_zeroes": true, 00:15:53.931 "zcopy": false, 00:15:53.931 "get_zone_info": false, 00:15:53.931 "zone_management": false, 00:15:53.931 "zone_append": false, 00:15:53.931 "compare": false, 00:15:53.931 "compare_and_write": false, 00:15:53.931 "abort": false, 00:15:53.931 "seek_hole": false, 00:15:53.931 "seek_data": false, 00:15:53.931 "copy": false, 00:15:53.931 "nvme_iov_md": false 00:15:53.931 }, 00:15:53.931 "memory_domains": [ 00:15:53.931 { 00:15:53.931 "dma_device_id": "system", 00:15:53.931 "dma_device_type": 1 00:15:53.931 }, 00:15:53.931 { 00:15:53.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:53.931 "dma_device_type": 2 00:15:53.931 }, 00:15:53.931 { 00:15:53.931 "dma_device_id": "system", 00:15:53.931 "dma_device_type": 1 00:15:53.931 }, 00:15:53.931 { 00:15:53.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:53.931 "dma_device_type": 2 00:15:53.931 } 00:15:53.931 ], 00:15:53.931 "driver_specific": { 00:15:53.931 "raid": { 00:15:53.931 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:53.931 "strip_size_kb": 0, 00:15:53.931 "state": "online", 00:15:53.931 "raid_level": "raid1", 00:15:53.931 "superblock": true, 00:15:53.931 "num_base_bdevs": 2, 00:15:53.931 "num_base_bdevs_discovered": 2, 00:15:53.931 "num_base_bdevs_operational": 2, 00:15:53.931 "base_bdevs_list": [ 00:15:53.931 { 00:15:53.931 "name": "pt1", 00:15:53.931 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:53.931 "is_configured": true, 00:15:53.931 "data_offset": 256, 00:15:53.931 "data_size": 7936 00:15:53.931 }, 00:15:53.931 { 00:15:53.931 "name": "pt2", 00:15:53.931 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:53.931 "is_configured": true, 00:15:53.931 "data_offset": 256, 00:15:53.931 "data_size": 7936 00:15:53.931 } 00:15:53.931 ] 00:15:53.931 } 00:15:53.931 } 00:15:53.931 }' 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:53.931 pt2' 00:15:53.931 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.191 [2024-09-30 14:14:58.712897] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 7e212a68-c9ac-4544-865c-daa2eb190c55 '!=' 7e212a68-c9ac-4544-865c-daa2eb190c55 ']' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.191 [2024-09-30 14:14:58.760641] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.191 "name": "raid_bdev1", 00:15:54.191 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:54.191 "strip_size_kb": 0, 00:15:54.191 "state": "online", 00:15:54.191 "raid_level": "raid1", 00:15:54.191 "superblock": true, 00:15:54.191 "num_base_bdevs": 2, 00:15:54.191 "num_base_bdevs_discovered": 1, 00:15:54.191 "num_base_bdevs_operational": 1, 00:15:54.191 "base_bdevs_list": [ 00:15:54.191 { 00:15:54.191 "name": null, 00:15:54.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.191 "is_configured": false, 00:15:54.191 "data_offset": 0, 00:15:54.191 "data_size": 7936 00:15:54.191 }, 00:15:54.191 { 00:15:54.191 "name": "pt2", 00:15:54.191 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:54.191 "is_configured": true, 00:15:54.191 "data_offset": 256, 00:15:54.191 "data_size": 7936 00:15:54.191 } 00:15:54.191 ] 00:15:54.191 }' 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.191 14:14:58 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.760 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:54.760 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.760 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.760 [2024-09-30 14:14:59.163911] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:54.760 [2024-09-30 14:14:59.163984] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:54.760 [2024-09-30 14:14:59.164049] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:54.760 [2024-09-30 14:14:59.164098] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:54.760 [2024-09-30 14:14:59.164132] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:15:54.760 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.760 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.761 [2024-09-30 14:14:59.219811] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:54.761 [2024-09-30 14:14:59.219899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:54.761 [2024-09-30 14:14:59.219929] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:15:54.761 [2024-09-30 14:14:59.219957] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:54.761 [2024-09-30 14:14:59.221815] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:54.761 [2024-09-30 14:14:59.221886] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:54.761 [2024-09-30 14:14:59.221943] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:54.761 [2024-09-30 14:14:59.221984] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:54.761 [2024-09-30 14:14:59.222069] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:15:54.761 [2024-09-30 14:14:59.222105] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:54.761 [2024-09-30 14:14:59.222204] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:54.761 [2024-09-30 14:14:59.222305] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:15:54.761 [2024-09-30 14:14:59.222338] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:15:54.761 [2024-09-30 14:14:59.222427] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:54.761 pt2 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.761 "name": "raid_bdev1", 00:15:54.761 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:54.761 "strip_size_kb": 0, 00:15:54.761 "state": "online", 00:15:54.761 "raid_level": "raid1", 00:15:54.761 "superblock": true, 00:15:54.761 "num_base_bdevs": 2, 00:15:54.761 "num_base_bdevs_discovered": 1, 00:15:54.761 "num_base_bdevs_operational": 1, 00:15:54.761 "base_bdevs_list": [ 00:15:54.761 { 00:15:54.761 "name": null, 00:15:54.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.761 "is_configured": false, 00:15:54.761 "data_offset": 256, 00:15:54.761 "data_size": 7936 00:15:54.761 }, 00:15:54.761 { 00:15:54.761 "name": "pt2", 00:15:54.761 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:54.761 "is_configured": true, 00:15:54.761 "data_offset": 256, 00:15:54.761 "data_size": 7936 00:15:54.761 } 00:15:54.761 ] 00:15:54.761 }' 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.761 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.330 [2024-09-30 14:14:59.691134] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:55.330 [2024-09-30 14:14:59.691157] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:55.330 [2024-09-30 14:14:59.691198] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:55.330 [2024-09-30 14:14:59.691232] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:55.330 [2024-09-30 14:14:59.691239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.330 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.330 [2024-09-30 14:14:59.755027] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:55.330 [2024-09-30 14:14:59.755071] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.330 [2024-09-30 14:14:59.755090] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:15:55.330 [2024-09-30 14:14:59.755098] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.330 [2024-09-30 14:14:59.756936] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.330 [2024-09-30 14:14:59.756971] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:55.330 [2024-09-30 14:14:59.757009] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:55.330 [2024-09-30 14:14:59.757037] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:55.330 [2024-09-30 14:14:59.757117] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:55.331 [2024-09-30 14:14:59.757127] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:55.331 [2024-09-30 14:14:59.757140] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:15:55.331 [2024-09-30 14:14:59.757174] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:55.331 [2024-09-30 14:14:59.757224] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:15:55.331 [2024-09-30 14:14:59.757239] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:55.331 [2024-09-30 14:14:59.757300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:55.331 [2024-09-30 14:14:59.757366] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:15:55.331 [2024-09-30 14:14:59.757376] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:15:55.331 [2024-09-30 14:14:59.757439] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.331 pt1 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.331 "name": "raid_bdev1", 00:15:55.331 "uuid": "7e212a68-c9ac-4544-865c-daa2eb190c55", 00:15:55.331 "strip_size_kb": 0, 00:15:55.331 "state": "online", 00:15:55.331 "raid_level": "raid1", 00:15:55.331 "superblock": true, 00:15:55.331 "num_base_bdevs": 2, 00:15:55.331 "num_base_bdevs_discovered": 1, 00:15:55.331 "num_base_bdevs_operational": 1, 00:15:55.331 "base_bdevs_list": [ 00:15:55.331 { 00:15:55.331 "name": null, 00:15:55.331 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.331 "is_configured": false, 00:15:55.331 "data_offset": 256, 00:15:55.331 "data_size": 7936 00:15:55.331 }, 00:15:55.331 { 00:15:55.331 "name": "pt2", 00:15:55.331 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:55.331 "is_configured": true, 00:15:55.331 "data_offset": 256, 00:15:55.331 "data_size": 7936 00:15:55.331 } 00:15:55.331 ] 00:15:55.331 }' 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.331 14:14:59 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.591 [2024-09-30 14:15:00.218412] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:55.591 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 7e212a68-c9ac-4544-865c-daa2eb190c55 '!=' 7e212a68-c9ac-4544-865c-daa2eb190c55 ']' 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 98800 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@950 -- # '[' -z 98800 ']' 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # kill -0 98800 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # uname 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98800 00:15:55.851 killing process with pid 98800 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98800' 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@969 -- # kill 98800 00:15:55.851 [2024-09-30 14:15:00.285775] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:55.851 [2024-09-30 14:15:00.285832] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:55.851 [2024-09-30 14:15:00.285862] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:55.851 [2024-09-30 14:15:00.285874] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:15:55.851 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@974 -- # wait 98800 00:15:55.851 [2024-09-30 14:15:00.310211] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:56.111 ************************************ 00:15:56.111 END TEST raid_superblock_test_md_separate 00:15:56.111 ************************************ 00:15:56.111 14:15:00 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:15:56.111 00:15:56.111 real 0m4.823s 00:15:56.111 user 0m7.794s 00:15:56.111 sys 0m1.062s 00:15:56.111 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:56.111 14:15:00 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.111 14:15:00 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:15:56.111 14:15:00 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:15:56.111 14:15:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:56.111 14:15:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:56.111 14:15:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:56.111 ************************************ 00:15:56.111 START TEST raid_rebuild_test_sb_md_separate 00:15:56.111 ************************************ 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=99111 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 99111 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 99111 ']' 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:56.111 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:56.111 14:15:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.111 [2024-09-30 14:15:00.745200] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:15:56.111 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:56.111 Zero copy mechanism will not be used. 00:15:56.111 [2024-09-30 14:15:00.745462] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99111 ] 00:15:56.370 [2024-09-30 14:15:00.883342] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:15:56.370 [2024-09-30 14:15:00.912707] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:56.370 [2024-09-30 14:15:00.959534] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:56.370 [2024-09-30 14:15:01.001542] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:56.370 [2024-09-30 14:15:01.001575] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.940 BaseBdev1_malloc 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.940 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 [2024-09-30 14:15:01.596269] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:57.201 [2024-09-30 14:15:01.596340] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.201 [2024-09-30 14:15:01.596365] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:57.201 [2024-09-30 14:15:01.596401] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.201 [2024-09-30 14:15:01.598311] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.201 [2024-09-30 14:15:01.598357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:57.201 BaseBdev1 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 BaseBdev2_malloc 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 [2024-09-30 14:15:01.638889] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:57.201 [2024-09-30 14:15:01.638979] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.201 [2024-09-30 14:15:01.639013] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:57.201 [2024-09-30 14:15:01.639034] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.201 [2024-09-30 14:15:01.642155] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.201 [2024-09-30 14:15:01.642286] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:57.201 BaseBdev2 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 spare_malloc 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 spare_delay 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 [2024-09-30 14:15:01.680236] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:57.201 [2024-09-30 14:15:01.680291] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.201 [2024-09-30 14:15:01.680314] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:57.201 [2024-09-30 14:15:01.680325] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.201 [2024-09-30 14:15:01.682194] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.201 [2024-09-30 14:15:01.682230] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:57.201 spare 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.201 [2024-09-30 14:15:01.692264] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:57.201 [2024-09-30 14:15:01.694036] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:57.201 [2024-09-30 14:15:01.694174] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:57.201 [2024-09-30 14:15:01.694188] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:57.201 [2024-09-30 14:15:01.694257] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:57.201 [2024-09-30 14:15:01.694354] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:57.201 [2024-09-30 14:15:01.694365] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:57.201 [2024-09-30 14:15:01.694445] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:57.201 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.202 "name": "raid_bdev1", 00:15:57.202 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:15:57.202 "strip_size_kb": 0, 00:15:57.202 "state": "online", 00:15:57.202 "raid_level": "raid1", 00:15:57.202 "superblock": true, 00:15:57.202 "num_base_bdevs": 2, 00:15:57.202 "num_base_bdevs_discovered": 2, 00:15:57.202 "num_base_bdevs_operational": 2, 00:15:57.202 "base_bdevs_list": [ 00:15:57.202 { 00:15:57.202 "name": "BaseBdev1", 00:15:57.202 "uuid": "b52a34a8-6e7e-5e9a-85f0-0a1514eed35f", 00:15:57.202 "is_configured": true, 00:15:57.202 "data_offset": 256, 00:15:57.202 "data_size": 7936 00:15:57.202 }, 00:15:57.202 { 00:15:57.202 "name": "BaseBdev2", 00:15:57.202 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:15:57.202 "is_configured": true, 00:15:57.202 "data_offset": 256, 00:15:57.202 "data_size": 7936 00:15:57.202 } 00:15:57.202 ] 00:15:57.202 }' 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.202 14:15:01 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.462 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:57.462 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.462 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:57.462 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.722 [2024-09-30 14:15:02.119830] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:57.722 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:57.722 [2024-09-30 14:15:02.367317] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:57.983 /dev/nbd0 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:57.983 1+0 records in 00:15:57.983 1+0 records out 00:15:57.983 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000469471 s, 8.7 MB/s 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:57.983 14:15:02 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:58.550 7936+0 records in 00:15:58.550 7936+0 records out 00:15:58.550 32505856 bytes (33 MB, 31 MiB) copied, 0.592206 s, 54.9 MB/s 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:58.550 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:58.809 [2024-09-30 14:15:03.233456] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.809 [2024-09-30 14:15:03.265463] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.809 "name": "raid_bdev1", 00:15:58.809 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:15:58.809 "strip_size_kb": 0, 00:15:58.809 "state": "online", 00:15:58.809 "raid_level": "raid1", 00:15:58.809 "superblock": true, 00:15:58.809 "num_base_bdevs": 2, 00:15:58.809 "num_base_bdevs_discovered": 1, 00:15:58.809 "num_base_bdevs_operational": 1, 00:15:58.809 "base_bdevs_list": [ 00:15:58.809 { 00:15:58.809 "name": null, 00:15:58.809 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.809 "is_configured": false, 00:15:58.809 "data_offset": 0, 00:15:58.809 "data_size": 7936 00:15:58.809 }, 00:15:58.809 { 00:15:58.809 "name": "BaseBdev2", 00:15:58.809 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:15:58.809 "is_configured": true, 00:15:58.809 "data_offset": 256, 00:15:58.809 "data_size": 7936 00:15:58.809 } 00:15:58.809 ] 00:15:58.809 }' 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.809 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.377 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:59.377 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.377 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.377 [2024-09-30 14:15:03.744665] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:59.377 [2024-09-30 14:15:03.746406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:15:59.377 [2024-09-30 14:15:03.748091] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:59.377 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.377 14:15:03 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.315 "name": "raid_bdev1", 00:16:00.315 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:00.315 "strip_size_kb": 0, 00:16:00.315 "state": "online", 00:16:00.315 "raid_level": "raid1", 00:16:00.315 "superblock": true, 00:16:00.315 "num_base_bdevs": 2, 00:16:00.315 "num_base_bdevs_discovered": 2, 00:16:00.315 "num_base_bdevs_operational": 2, 00:16:00.315 "process": { 00:16:00.315 "type": "rebuild", 00:16:00.315 "target": "spare", 00:16:00.315 "progress": { 00:16:00.315 "blocks": 2560, 00:16:00.315 "percent": 32 00:16:00.315 } 00:16:00.315 }, 00:16:00.315 "base_bdevs_list": [ 00:16:00.315 { 00:16:00.315 "name": "spare", 00:16:00.315 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:00.315 "is_configured": true, 00:16:00.315 "data_offset": 256, 00:16:00.315 "data_size": 7936 00:16:00.315 }, 00:16:00.315 { 00:16:00.315 "name": "BaseBdev2", 00:16:00.315 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:00.315 "is_configured": true, 00:16:00.315 "data_offset": 256, 00:16:00.315 "data_size": 7936 00:16:00.315 } 00:16:00.315 ] 00:16:00.315 }' 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.315 [2024-09-30 14:15:04.887283] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:00.315 [2024-09-30 14:15:04.952599] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:00.315 [2024-09-30 14:15:04.952668] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:00.315 [2024-09-30 14:15:04.952684] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:00.315 [2024-09-30 14:15:04.952694] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:00.315 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.316 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.575 14:15:04 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.575 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.575 "name": "raid_bdev1", 00:16:00.575 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:00.575 "strip_size_kb": 0, 00:16:00.575 "state": "online", 00:16:00.575 "raid_level": "raid1", 00:16:00.575 "superblock": true, 00:16:00.575 "num_base_bdevs": 2, 00:16:00.575 "num_base_bdevs_discovered": 1, 00:16:00.575 "num_base_bdevs_operational": 1, 00:16:00.575 "base_bdevs_list": [ 00:16:00.575 { 00:16:00.575 "name": null, 00:16:00.575 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.575 "is_configured": false, 00:16:00.575 "data_offset": 0, 00:16:00.575 "data_size": 7936 00:16:00.575 }, 00:16:00.575 { 00:16:00.575 "name": "BaseBdev2", 00:16:00.575 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:00.575 "is_configured": true, 00:16:00.575 "data_offset": 256, 00:16:00.575 "data_size": 7936 00:16:00.575 } 00:16:00.575 ] 00:16:00.575 }' 00:16:00.575 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.575 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.834 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.835 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.835 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.835 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.835 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.835 "name": "raid_bdev1", 00:16:00.835 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:00.835 "strip_size_kb": 0, 00:16:00.835 "state": "online", 00:16:00.835 "raid_level": "raid1", 00:16:00.835 "superblock": true, 00:16:00.835 "num_base_bdevs": 2, 00:16:00.835 "num_base_bdevs_discovered": 1, 00:16:00.835 "num_base_bdevs_operational": 1, 00:16:00.835 "base_bdevs_list": [ 00:16:00.835 { 00:16:00.835 "name": null, 00:16:00.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.835 "is_configured": false, 00:16:00.835 "data_offset": 0, 00:16:00.835 "data_size": 7936 00:16:00.835 }, 00:16:00.835 { 00:16:00.835 "name": "BaseBdev2", 00:16:00.835 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:00.835 "is_configured": true, 00:16:00.835 "data_offset": 256, 00:16:00.835 "data_size": 7936 00:16:00.835 } 00:16:00.835 ] 00:16:00.835 }' 00:16:00.835 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.094 [2024-09-30 14:15:05.574296] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:01.094 [2024-09-30 14:15:05.575774] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d400 00:16:01.094 [2024-09-30 14:15:05.577575] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.094 14:15:05 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:02.033 "name": "raid_bdev1", 00:16:02.033 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:02.033 "strip_size_kb": 0, 00:16:02.033 "state": "online", 00:16:02.033 "raid_level": "raid1", 00:16:02.033 "superblock": true, 00:16:02.033 "num_base_bdevs": 2, 00:16:02.033 "num_base_bdevs_discovered": 2, 00:16:02.033 "num_base_bdevs_operational": 2, 00:16:02.033 "process": { 00:16:02.033 "type": "rebuild", 00:16:02.033 "target": "spare", 00:16:02.033 "progress": { 00:16:02.033 "blocks": 2560, 00:16:02.033 "percent": 32 00:16:02.033 } 00:16:02.033 }, 00:16:02.033 "base_bdevs_list": [ 00:16:02.033 { 00:16:02.033 "name": "spare", 00:16:02.033 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:02.033 "is_configured": true, 00:16:02.033 "data_offset": 256, 00:16:02.033 "data_size": 7936 00:16:02.033 }, 00:16:02.033 { 00:16:02.033 "name": "BaseBdev2", 00:16:02.033 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:02.033 "is_configured": true, 00:16:02.033 "data_offset": 256, 00:16:02.033 "data_size": 7936 00:16:02.033 } 00:16:02.033 ] 00:16:02.033 }' 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:02.033 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:02.292 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=587 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.292 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:02.292 "name": "raid_bdev1", 00:16:02.293 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:02.293 "strip_size_kb": 0, 00:16:02.293 "state": "online", 00:16:02.293 "raid_level": "raid1", 00:16:02.293 "superblock": true, 00:16:02.293 "num_base_bdevs": 2, 00:16:02.293 "num_base_bdevs_discovered": 2, 00:16:02.293 "num_base_bdevs_operational": 2, 00:16:02.293 "process": { 00:16:02.293 "type": "rebuild", 00:16:02.293 "target": "spare", 00:16:02.293 "progress": { 00:16:02.293 "blocks": 2816, 00:16:02.293 "percent": 35 00:16:02.293 } 00:16:02.293 }, 00:16:02.293 "base_bdevs_list": [ 00:16:02.293 { 00:16:02.293 "name": "spare", 00:16:02.293 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:02.293 "is_configured": true, 00:16:02.293 "data_offset": 256, 00:16:02.293 "data_size": 7936 00:16:02.293 }, 00:16:02.293 { 00:16:02.293 "name": "BaseBdev2", 00:16:02.293 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:02.293 "is_configured": true, 00:16:02.293 "data_offset": 256, 00:16:02.293 "data_size": 7936 00:16:02.293 } 00:16:02.293 ] 00:16:02.293 }' 00:16:02.293 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:02.293 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:02.293 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:02.293 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:02.293 14:15:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.230 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.490 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:03.490 "name": "raid_bdev1", 00:16:03.490 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:03.490 "strip_size_kb": 0, 00:16:03.490 "state": "online", 00:16:03.490 "raid_level": "raid1", 00:16:03.490 "superblock": true, 00:16:03.490 "num_base_bdevs": 2, 00:16:03.490 "num_base_bdevs_discovered": 2, 00:16:03.490 "num_base_bdevs_operational": 2, 00:16:03.490 "process": { 00:16:03.490 "type": "rebuild", 00:16:03.490 "target": "spare", 00:16:03.490 "progress": { 00:16:03.490 "blocks": 5632, 00:16:03.490 "percent": 70 00:16:03.490 } 00:16:03.490 }, 00:16:03.490 "base_bdevs_list": [ 00:16:03.490 { 00:16:03.490 "name": "spare", 00:16:03.490 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:03.490 "is_configured": true, 00:16:03.490 "data_offset": 256, 00:16:03.490 "data_size": 7936 00:16:03.490 }, 00:16:03.490 { 00:16:03.490 "name": "BaseBdev2", 00:16:03.490 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:03.490 "is_configured": true, 00:16:03.490 "data_offset": 256, 00:16:03.490 "data_size": 7936 00:16:03.490 } 00:16:03.490 ] 00:16:03.490 }' 00:16:03.490 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:03.490 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:03.490 14:15:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:03.490 14:15:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:03.490 14:15:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:04.059 [2024-09-30 14:15:08.687569] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:04.059 [2024-09-30 14:15:08.687685] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:04.059 [2024-09-30 14:15:08.687798] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:04.628 "name": "raid_bdev1", 00:16:04.628 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:04.628 "strip_size_kb": 0, 00:16:04.628 "state": "online", 00:16:04.628 "raid_level": "raid1", 00:16:04.628 "superblock": true, 00:16:04.628 "num_base_bdevs": 2, 00:16:04.628 "num_base_bdevs_discovered": 2, 00:16:04.628 "num_base_bdevs_operational": 2, 00:16:04.628 "base_bdevs_list": [ 00:16:04.628 { 00:16:04.628 "name": "spare", 00:16:04.628 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:04.628 "is_configured": true, 00:16:04.628 "data_offset": 256, 00:16:04.628 "data_size": 7936 00:16:04.628 }, 00:16:04.628 { 00:16:04.628 "name": "BaseBdev2", 00:16:04.628 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:04.628 "is_configured": true, 00:16:04.628 "data_offset": 256, 00:16:04.628 "data_size": 7936 00:16:04.628 } 00:16:04.628 ] 00:16:04.628 }' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:04.628 "name": "raid_bdev1", 00:16:04.628 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:04.628 "strip_size_kb": 0, 00:16:04.628 "state": "online", 00:16:04.628 "raid_level": "raid1", 00:16:04.628 "superblock": true, 00:16:04.628 "num_base_bdevs": 2, 00:16:04.628 "num_base_bdevs_discovered": 2, 00:16:04.628 "num_base_bdevs_operational": 2, 00:16:04.628 "base_bdevs_list": [ 00:16:04.628 { 00:16:04.628 "name": "spare", 00:16:04.628 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:04.628 "is_configured": true, 00:16:04.628 "data_offset": 256, 00:16:04.628 "data_size": 7936 00:16:04.628 }, 00:16:04.628 { 00:16:04.628 "name": "BaseBdev2", 00:16:04.628 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:04.628 "is_configured": true, 00:16:04.628 "data_offset": 256, 00:16:04.628 "data_size": 7936 00:16:04.628 } 00:16:04.628 ] 00:16:04.628 }' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:04.628 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.887 "name": "raid_bdev1", 00:16:04.887 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:04.887 "strip_size_kb": 0, 00:16:04.887 "state": "online", 00:16:04.887 "raid_level": "raid1", 00:16:04.887 "superblock": true, 00:16:04.887 "num_base_bdevs": 2, 00:16:04.887 "num_base_bdevs_discovered": 2, 00:16:04.887 "num_base_bdevs_operational": 2, 00:16:04.887 "base_bdevs_list": [ 00:16:04.887 { 00:16:04.887 "name": "spare", 00:16:04.887 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:04.887 "is_configured": true, 00:16:04.887 "data_offset": 256, 00:16:04.887 "data_size": 7936 00:16:04.887 }, 00:16:04.887 { 00:16:04.887 "name": "BaseBdev2", 00:16:04.887 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:04.887 "is_configured": true, 00:16:04.887 "data_offset": 256, 00:16:04.887 "data_size": 7936 00:16:04.887 } 00:16:04.887 ] 00:16:04.887 }' 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.887 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.147 [2024-09-30 14:15:09.773418] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:05.147 [2024-09-30 14:15:09.773447] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:05.147 [2024-09-30 14:15:09.773533] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:05.147 [2024-09-30 14:15:09.773591] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:05.147 [2024-09-30 14:15:09.773600] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.147 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:05.406 14:15:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:05.406 /dev/nbd0 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:05.666 1+0 records in 00:16:05.666 1+0 records out 00:16:05.666 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000217562 s, 18.8 MB/s 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:05.666 /dev/nbd1 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:05.666 1+0 records in 00:16:05.666 1+0 records out 00:16:05.666 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000324344 s, 12.6 MB/s 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:05.666 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:05.927 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.187 [2024-09-30 14:15:10.832588] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:06.187 [2024-09-30 14:15:10.832646] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.187 [2024-09-30 14:15:10.832673] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:06.187 [2024-09-30 14:15:10.832682] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.187 [2024-09-30 14:15:10.835201] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.187 [2024-09-30 14:15:10.835239] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:06.187 [2024-09-30 14:15:10.835294] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:06.187 [2024-09-30 14:15:10.835340] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:06.187 [2024-09-30 14:15:10.835463] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:06.187 spare 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.187 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.447 [2024-09-30 14:15:10.935384] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:16:06.447 [2024-09-30 14:15:10.935501] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:06.447 [2024-09-30 14:15:10.935607] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:16:06.447 [2024-09-30 14:15:10.935728] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:16:06.447 [2024-09-30 14:15:10.935738] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:16:06.447 [2024-09-30 14:15:10.935825] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.447 "name": "raid_bdev1", 00:16:06.447 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:06.447 "strip_size_kb": 0, 00:16:06.447 "state": "online", 00:16:06.447 "raid_level": "raid1", 00:16:06.447 "superblock": true, 00:16:06.447 "num_base_bdevs": 2, 00:16:06.447 "num_base_bdevs_discovered": 2, 00:16:06.447 "num_base_bdevs_operational": 2, 00:16:06.447 "base_bdevs_list": [ 00:16:06.447 { 00:16:06.447 "name": "spare", 00:16:06.447 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:06.447 "is_configured": true, 00:16:06.447 "data_offset": 256, 00:16:06.447 "data_size": 7936 00:16:06.447 }, 00:16:06.447 { 00:16:06.447 "name": "BaseBdev2", 00:16:06.447 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:06.447 "is_configured": true, 00:16:06.447 "data_offset": 256, 00:16:06.447 "data_size": 7936 00:16:06.447 } 00:16:06.447 ] 00:16:06.447 }' 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.447 14:15:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.707 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:06.967 "name": "raid_bdev1", 00:16:06.967 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:06.967 "strip_size_kb": 0, 00:16:06.967 "state": "online", 00:16:06.967 "raid_level": "raid1", 00:16:06.967 "superblock": true, 00:16:06.967 "num_base_bdevs": 2, 00:16:06.967 "num_base_bdevs_discovered": 2, 00:16:06.967 "num_base_bdevs_operational": 2, 00:16:06.967 "base_bdevs_list": [ 00:16:06.967 { 00:16:06.967 "name": "spare", 00:16:06.967 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:06.967 "is_configured": true, 00:16:06.967 "data_offset": 256, 00:16:06.967 "data_size": 7936 00:16:06.967 }, 00:16:06.967 { 00:16:06.967 "name": "BaseBdev2", 00:16:06.967 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:06.967 "is_configured": true, 00:16:06.967 "data_offset": 256, 00:16:06.967 "data_size": 7936 00:16:06.967 } 00:16:06.967 ] 00:16:06.967 }' 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.967 [2024-09-30 14:15:11.515584] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.967 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.967 "name": "raid_bdev1", 00:16:06.967 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:06.967 "strip_size_kb": 0, 00:16:06.967 "state": "online", 00:16:06.967 "raid_level": "raid1", 00:16:06.968 "superblock": true, 00:16:06.968 "num_base_bdevs": 2, 00:16:06.968 "num_base_bdevs_discovered": 1, 00:16:06.968 "num_base_bdevs_operational": 1, 00:16:06.968 "base_bdevs_list": [ 00:16:06.968 { 00:16:06.968 "name": null, 00:16:06.968 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.968 "is_configured": false, 00:16:06.968 "data_offset": 0, 00:16:06.968 "data_size": 7936 00:16:06.968 }, 00:16:06.968 { 00:16:06.968 "name": "BaseBdev2", 00:16:06.968 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:06.968 "is_configured": true, 00:16:06.968 "data_offset": 256, 00:16:06.968 "data_size": 7936 00:16:06.968 } 00:16:06.968 ] 00:16:06.968 }' 00:16:06.968 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.968 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.536 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:07.537 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.537 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.537 [2024-09-30 14:15:11.938972] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:07.537 [2024-09-30 14:15:11.939153] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:07.537 [2024-09-30 14:15:11.939232] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:07.537 [2024-09-30 14:15:11.939289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:07.537 [2024-09-30 14:15:11.940928] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:16:07.537 [2024-09-30 14:15:11.942716] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:07.537 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.537 14:15:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.475 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.475 "name": "raid_bdev1", 00:16:08.475 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:08.475 "strip_size_kb": 0, 00:16:08.475 "state": "online", 00:16:08.475 "raid_level": "raid1", 00:16:08.475 "superblock": true, 00:16:08.475 "num_base_bdevs": 2, 00:16:08.475 "num_base_bdevs_discovered": 2, 00:16:08.475 "num_base_bdevs_operational": 2, 00:16:08.475 "process": { 00:16:08.475 "type": "rebuild", 00:16:08.475 "target": "spare", 00:16:08.475 "progress": { 00:16:08.475 "blocks": 2560, 00:16:08.475 "percent": 32 00:16:08.475 } 00:16:08.475 }, 00:16:08.475 "base_bdevs_list": [ 00:16:08.475 { 00:16:08.476 "name": "spare", 00:16:08.476 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:08.476 "is_configured": true, 00:16:08.476 "data_offset": 256, 00:16:08.476 "data_size": 7936 00:16:08.476 }, 00:16:08.476 { 00:16:08.476 "name": "BaseBdev2", 00:16:08.476 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:08.476 "is_configured": true, 00:16:08.476 "data_offset": 256, 00:16:08.476 "data_size": 7936 00:16:08.476 } 00:16:08.476 ] 00:16:08.476 }' 00:16:08.476 14:15:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.476 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.476 [2024-09-30 14:15:13.089941] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.736 [2024-09-30 14:15:13.146726] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:08.736 [2024-09-30 14:15:13.146781] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.736 [2024-09-30 14:15:13.146794] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.736 [2024-09-30 14:15:13.146802] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.736 "name": "raid_bdev1", 00:16:08.736 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:08.736 "strip_size_kb": 0, 00:16:08.736 "state": "online", 00:16:08.736 "raid_level": "raid1", 00:16:08.736 "superblock": true, 00:16:08.736 "num_base_bdevs": 2, 00:16:08.736 "num_base_bdevs_discovered": 1, 00:16:08.736 "num_base_bdevs_operational": 1, 00:16:08.736 "base_bdevs_list": [ 00:16:08.736 { 00:16:08.736 "name": null, 00:16:08.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.736 "is_configured": false, 00:16:08.736 "data_offset": 0, 00:16:08.736 "data_size": 7936 00:16:08.736 }, 00:16:08.736 { 00:16:08.736 "name": "BaseBdev2", 00:16:08.736 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:08.736 "is_configured": true, 00:16:08.736 "data_offset": 256, 00:16:08.736 "data_size": 7936 00:16:08.736 } 00:16:08.736 ] 00:16:08.736 }' 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.736 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.996 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:08.996 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.996 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.996 [2024-09-30 14:15:13.561716] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:08.996 [2024-09-30 14:15:13.561831] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.996 [2024-09-30 14:15:13.561872] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:08.996 [2024-09-30 14:15:13.561903] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.996 [2024-09-30 14:15:13.562125] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.996 [2024-09-30 14:15:13.562176] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:08.996 [2024-09-30 14:15:13.562246] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:08.996 [2024-09-30 14:15:13.562286] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:08.996 [2024-09-30 14:15:13.562328] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:08.996 [2024-09-30 14:15:13.562390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:08.996 [2024-09-30 14:15:13.563654] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1dc0 00:16:08.996 [2024-09-30 14:15:13.565387] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:08.996 spare 00:16:08.996 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.996 14:15:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.934 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.195 "name": "raid_bdev1", 00:16:10.195 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:10.195 "strip_size_kb": 0, 00:16:10.195 "state": "online", 00:16:10.195 "raid_level": "raid1", 00:16:10.195 "superblock": true, 00:16:10.195 "num_base_bdevs": 2, 00:16:10.195 "num_base_bdevs_discovered": 2, 00:16:10.195 "num_base_bdevs_operational": 2, 00:16:10.195 "process": { 00:16:10.195 "type": "rebuild", 00:16:10.195 "target": "spare", 00:16:10.195 "progress": { 00:16:10.195 "blocks": 2560, 00:16:10.195 "percent": 32 00:16:10.195 } 00:16:10.195 }, 00:16:10.195 "base_bdevs_list": [ 00:16:10.195 { 00:16:10.195 "name": "spare", 00:16:10.195 "uuid": "b6f10c8b-7ff9-5e5f-9246-fa9c5310189a", 00:16:10.195 "is_configured": true, 00:16:10.195 "data_offset": 256, 00:16:10.195 "data_size": 7936 00:16:10.195 }, 00:16:10.195 { 00:16:10.195 "name": "BaseBdev2", 00:16:10.195 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:10.195 "is_configured": true, 00:16:10.195 "data_offset": 256, 00:16:10.195 "data_size": 7936 00:16:10.195 } 00:16:10.195 ] 00:16:10.195 }' 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.195 [2024-09-30 14:15:14.732350] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:10.195 [2024-09-30 14:15:14.769125] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:10.195 [2024-09-30 14:15:14.769173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:10.195 [2024-09-30 14:15:14.769189] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:10.195 [2024-09-30 14:15:14.769196] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:10.195 "name": "raid_bdev1", 00:16:10.195 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:10.195 "strip_size_kb": 0, 00:16:10.195 "state": "online", 00:16:10.195 "raid_level": "raid1", 00:16:10.195 "superblock": true, 00:16:10.195 "num_base_bdevs": 2, 00:16:10.195 "num_base_bdevs_discovered": 1, 00:16:10.195 "num_base_bdevs_operational": 1, 00:16:10.195 "base_bdevs_list": [ 00:16:10.195 { 00:16:10.195 "name": null, 00:16:10.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.195 "is_configured": false, 00:16:10.195 "data_offset": 0, 00:16:10.195 "data_size": 7936 00:16:10.195 }, 00:16:10.195 { 00:16:10.195 "name": "BaseBdev2", 00:16:10.195 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:10.195 "is_configured": true, 00:16:10.195 "data_offset": 256, 00:16:10.195 "data_size": 7936 00:16:10.195 } 00:16:10.195 ] 00:16:10.195 }' 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:10.195 14:15:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.765 "name": "raid_bdev1", 00:16:10.765 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:10.765 "strip_size_kb": 0, 00:16:10.765 "state": "online", 00:16:10.765 "raid_level": "raid1", 00:16:10.765 "superblock": true, 00:16:10.765 "num_base_bdevs": 2, 00:16:10.765 "num_base_bdevs_discovered": 1, 00:16:10.765 "num_base_bdevs_operational": 1, 00:16:10.765 "base_bdevs_list": [ 00:16:10.765 { 00:16:10.765 "name": null, 00:16:10.765 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.765 "is_configured": false, 00:16:10.765 "data_offset": 0, 00:16:10.765 "data_size": 7936 00:16:10.765 }, 00:16:10.765 { 00:16:10.765 "name": "BaseBdev2", 00:16:10.765 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:10.765 "is_configured": true, 00:16:10.765 "data_offset": 256, 00:16:10.765 "data_size": 7936 00:16:10.765 } 00:16:10.765 ] 00:16:10.765 }' 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:10.765 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.766 [2024-09-30 14:15:15.319873] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:10.766 [2024-09-30 14:15:15.319968] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:10.766 [2024-09-30 14:15:15.319993] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:10.766 [2024-09-30 14:15:15.320003] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:10.766 [2024-09-30 14:15:15.320166] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:10.766 [2024-09-30 14:15:15.320178] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:10.766 [2024-09-30 14:15:15.320225] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:10.766 [2024-09-30 14:15:15.320242] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:10.766 [2024-09-30 14:15:15.320251] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:10.766 [2024-09-30 14:15:15.320271] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:10.766 BaseBdev1 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.766 14:15:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:11.703 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.704 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.963 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.963 "name": "raid_bdev1", 00:16:11.963 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:11.963 "strip_size_kb": 0, 00:16:11.963 "state": "online", 00:16:11.963 "raid_level": "raid1", 00:16:11.963 "superblock": true, 00:16:11.963 "num_base_bdevs": 2, 00:16:11.963 "num_base_bdevs_discovered": 1, 00:16:11.963 "num_base_bdevs_operational": 1, 00:16:11.963 "base_bdevs_list": [ 00:16:11.963 { 00:16:11.963 "name": null, 00:16:11.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.963 "is_configured": false, 00:16:11.963 "data_offset": 0, 00:16:11.963 "data_size": 7936 00:16:11.963 }, 00:16:11.963 { 00:16:11.963 "name": "BaseBdev2", 00:16:11.963 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:11.963 "is_configured": true, 00:16:11.963 "data_offset": 256, 00:16:11.963 "data_size": 7936 00:16:11.963 } 00:16:11.963 ] 00:16:11.963 }' 00:16:11.963 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.963 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:12.223 "name": "raid_bdev1", 00:16:12.223 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:12.223 "strip_size_kb": 0, 00:16:12.223 "state": "online", 00:16:12.223 "raid_level": "raid1", 00:16:12.223 "superblock": true, 00:16:12.223 "num_base_bdevs": 2, 00:16:12.223 "num_base_bdevs_discovered": 1, 00:16:12.223 "num_base_bdevs_operational": 1, 00:16:12.223 "base_bdevs_list": [ 00:16:12.223 { 00:16:12.223 "name": null, 00:16:12.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.223 "is_configured": false, 00:16:12.223 "data_offset": 0, 00:16:12.223 "data_size": 7936 00:16:12.223 }, 00:16:12.223 { 00:16:12.223 "name": "BaseBdev2", 00:16:12.223 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:12.223 "is_configured": true, 00:16:12.223 "data_offset": 256, 00:16:12.223 "data_size": 7936 00:16:12.223 } 00:16:12.223 ] 00:16:12.223 }' 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:12.223 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.224 [2024-09-30 14:15:16.861560] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:12.224 [2024-09-30 14:15:16.861682] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:12.224 [2024-09-30 14:15:16.861696] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:12.224 request: 00:16:12.224 { 00:16:12.224 "base_bdev": "BaseBdev1", 00:16:12.224 "raid_bdev": "raid_bdev1", 00:16:12.224 "method": "bdev_raid_add_base_bdev", 00:16:12.224 "req_id": 1 00:16:12.224 } 00:16:12.224 Got JSON-RPC error response 00:16:12.224 response: 00:16:12.224 { 00:16:12.224 "code": -22, 00:16:12.224 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:12.224 } 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:12.224 14:15:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.604 "name": "raid_bdev1", 00:16:13.604 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:13.604 "strip_size_kb": 0, 00:16:13.604 "state": "online", 00:16:13.604 "raid_level": "raid1", 00:16:13.604 "superblock": true, 00:16:13.604 "num_base_bdevs": 2, 00:16:13.604 "num_base_bdevs_discovered": 1, 00:16:13.604 "num_base_bdevs_operational": 1, 00:16:13.604 "base_bdevs_list": [ 00:16:13.604 { 00:16:13.604 "name": null, 00:16:13.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.604 "is_configured": false, 00:16:13.604 "data_offset": 0, 00:16:13.604 "data_size": 7936 00:16:13.604 }, 00:16:13.604 { 00:16:13.604 "name": "BaseBdev2", 00:16:13.604 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:13.604 "is_configured": true, 00:16:13.604 "data_offset": 256, 00:16:13.604 "data_size": 7936 00:16:13.604 } 00:16:13.604 ] 00:16:13.604 }' 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.604 14:15:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.864 "name": "raid_bdev1", 00:16:13.864 "uuid": "2735dcc7-7357-460e-bc58-8f7aaaa517c9", 00:16:13.864 "strip_size_kb": 0, 00:16:13.864 "state": "online", 00:16:13.864 "raid_level": "raid1", 00:16:13.864 "superblock": true, 00:16:13.864 "num_base_bdevs": 2, 00:16:13.864 "num_base_bdevs_discovered": 1, 00:16:13.864 "num_base_bdevs_operational": 1, 00:16:13.864 "base_bdevs_list": [ 00:16:13.864 { 00:16:13.864 "name": null, 00:16:13.864 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.864 "is_configured": false, 00:16:13.864 "data_offset": 0, 00:16:13.864 "data_size": 7936 00:16:13.864 }, 00:16:13.864 { 00:16:13.864 "name": "BaseBdev2", 00:16:13.864 "uuid": "fa15fd54-d438-58ea-b19d-0f8062e1995e", 00:16:13.864 "is_configured": true, 00:16:13.864 "data_offset": 256, 00:16:13.864 "data_size": 7936 00:16:13.864 } 00:16:13.864 ] 00:16:13.864 }' 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:13.864 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 99111 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 99111 ']' 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 99111 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99111 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99111' 00:16:13.865 killing process with pid 99111 00:16:13.865 Received shutdown signal, test time was about 60.000000 seconds 00:16:13.865 00:16:13.865 Latency(us) 00:16:13.865 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:13.865 =================================================================================================================== 00:16:13.865 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 99111 00:16:13.865 [2024-09-30 14:15:18.501333] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:13.865 [2024-09-30 14:15:18.501428] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:13.865 [2024-09-30 14:15:18.501466] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:13.865 [2024-09-30 14:15:18.501477] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:16:13.865 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 99111 00:16:14.125 [2024-09-30 14:15:18.535096] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:14.125 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:14.125 00:16:14.125 real 0m18.125s 00:16:14.125 user 0m23.935s 00:16:14.125 sys 0m2.647s 00:16:14.125 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:14.125 14:15:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.125 ************************************ 00:16:14.125 END TEST raid_rebuild_test_sb_md_separate 00:16:14.125 ************************************ 00:16:14.386 14:15:18 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:14.386 14:15:18 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:14.386 14:15:18 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:14.386 14:15:18 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:14.386 14:15:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:14.386 ************************************ 00:16:14.386 START TEST raid_state_function_test_sb_md_interleaved 00:16:14.386 ************************************ 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=99786 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 99786' 00:16:14.386 Process raid pid: 99786 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 99786 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 99786 ']' 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:14.386 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:14.386 14:15:18 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.386 [2024-09-30 14:15:18.950192] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:14.386 [2024-09-30 14:15:18.950453] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:14.645 [2024-09-30 14:15:19.090799] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:14.645 [2024-09-30 14:15:19.120143] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:14.645 [2024-09-30 14:15:19.166984] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:14.645 [2024-09-30 14:15:19.209217] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:14.645 [2024-09-30 14:15:19.209253] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:15.214 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:15.214 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.215 [2024-09-30 14:15:19.762580] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:15.215 [2024-09-30 14:15:19.762625] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:15.215 [2024-09-30 14:15:19.762643] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:15.215 [2024-09-30 14:15:19.762650] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:15.215 "name": "Existed_Raid", 00:16:15.215 "uuid": "49f0d4ee-ec48-4a18-ba59-b7f1ab314432", 00:16:15.215 "strip_size_kb": 0, 00:16:15.215 "state": "configuring", 00:16:15.215 "raid_level": "raid1", 00:16:15.215 "superblock": true, 00:16:15.215 "num_base_bdevs": 2, 00:16:15.215 "num_base_bdevs_discovered": 0, 00:16:15.215 "num_base_bdevs_operational": 2, 00:16:15.215 "base_bdevs_list": [ 00:16:15.215 { 00:16:15.215 "name": "BaseBdev1", 00:16:15.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.215 "is_configured": false, 00:16:15.215 "data_offset": 0, 00:16:15.215 "data_size": 0 00:16:15.215 }, 00:16:15.215 { 00:16:15.215 "name": "BaseBdev2", 00:16:15.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.215 "is_configured": false, 00:16:15.215 "data_offset": 0, 00:16:15.215 "data_size": 0 00:16:15.215 } 00:16:15.215 ] 00:16:15.215 }' 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:15.215 14:15:19 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 [2024-09-30 14:15:20.185733] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:15.785 [2024-09-30 14:15:20.185825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 [2024-09-30 14:15:20.197719] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:15.785 [2024-09-30 14:15:20.197795] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:15.785 [2024-09-30 14:15:20.197822] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:15.785 [2024-09-30 14:15:20.197842] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 [2024-09-30 14:15:20.218520] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:15.785 BaseBdev1 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 [ 00:16:15.785 { 00:16:15.785 "name": "BaseBdev1", 00:16:15.785 "aliases": [ 00:16:15.785 "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460" 00:16:15.785 ], 00:16:15.785 "product_name": "Malloc disk", 00:16:15.785 "block_size": 4128, 00:16:15.785 "num_blocks": 8192, 00:16:15.785 "uuid": "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460", 00:16:15.785 "md_size": 32, 00:16:15.785 "md_interleave": true, 00:16:15.785 "dif_type": 0, 00:16:15.785 "assigned_rate_limits": { 00:16:15.785 "rw_ios_per_sec": 0, 00:16:15.785 "rw_mbytes_per_sec": 0, 00:16:15.785 "r_mbytes_per_sec": 0, 00:16:15.785 "w_mbytes_per_sec": 0 00:16:15.785 }, 00:16:15.785 "claimed": true, 00:16:15.785 "claim_type": "exclusive_write", 00:16:15.785 "zoned": false, 00:16:15.785 "supported_io_types": { 00:16:15.785 "read": true, 00:16:15.785 "write": true, 00:16:15.785 "unmap": true, 00:16:15.785 "flush": true, 00:16:15.785 "reset": true, 00:16:15.785 "nvme_admin": false, 00:16:15.785 "nvme_io": false, 00:16:15.785 "nvme_io_md": false, 00:16:15.785 "write_zeroes": true, 00:16:15.785 "zcopy": true, 00:16:15.785 "get_zone_info": false, 00:16:15.785 "zone_management": false, 00:16:15.785 "zone_append": false, 00:16:15.785 "compare": false, 00:16:15.785 "compare_and_write": false, 00:16:15.785 "abort": true, 00:16:15.785 "seek_hole": false, 00:16:15.785 "seek_data": false, 00:16:15.785 "copy": true, 00:16:15.785 "nvme_iov_md": false 00:16:15.785 }, 00:16:15.785 "memory_domains": [ 00:16:15.785 { 00:16:15.785 "dma_device_id": "system", 00:16:15.785 "dma_device_type": 1 00:16:15.785 }, 00:16:15.785 { 00:16:15.785 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:15.785 "dma_device_type": 2 00:16:15.785 } 00:16:15.785 ], 00:16:15.785 "driver_specific": {} 00:16:15.785 } 00:16:15.785 ] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:15.785 "name": "Existed_Raid", 00:16:15.785 "uuid": "c7d8a935-6b39-4ac8-bcad-9954c1cee540", 00:16:15.785 "strip_size_kb": 0, 00:16:15.785 "state": "configuring", 00:16:15.785 "raid_level": "raid1", 00:16:15.785 "superblock": true, 00:16:15.785 "num_base_bdevs": 2, 00:16:15.785 "num_base_bdevs_discovered": 1, 00:16:15.785 "num_base_bdevs_operational": 2, 00:16:15.785 "base_bdevs_list": [ 00:16:15.785 { 00:16:15.785 "name": "BaseBdev1", 00:16:15.785 "uuid": "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460", 00:16:15.785 "is_configured": true, 00:16:15.785 "data_offset": 256, 00:16:15.785 "data_size": 7936 00:16:15.785 }, 00:16:15.785 { 00:16:15.785 "name": "BaseBdev2", 00:16:15.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.785 "is_configured": false, 00:16:15.785 "data_offset": 0, 00:16:15.785 "data_size": 0 00:16:15.785 } 00:16:15.785 ] 00:16:15.785 }' 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:15.785 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.045 [2024-09-30 14:15:20.657775] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:16.045 [2024-09-30 14:15:20.657874] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.045 [2024-09-30 14:15:20.665829] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:16.045 [2024-09-30 14:15:20.667534] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:16.045 [2024-09-30 14:15:20.667565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.045 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.305 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.305 "name": "Existed_Raid", 00:16:16.305 "uuid": "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1", 00:16:16.305 "strip_size_kb": 0, 00:16:16.305 "state": "configuring", 00:16:16.305 "raid_level": "raid1", 00:16:16.305 "superblock": true, 00:16:16.305 "num_base_bdevs": 2, 00:16:16.305 "num_base_bdevs_discovered": 1, 00:16:16.305 "num_base_bdevs_operational": 2, 00:16:16.305 "base_bdevs_list": [ 00:16:16.305 { 00:16:16.305 "name": "BaseBdev1", 00:16:16.305 "uuid": "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460", 00:16:16.305 "is_configured": true, 00:16:16.305 "data_offset": 256, 00:16:16.305 "data_size": 7936 00:16:16.305 }, 00:16:16.305 { 00:16:16.305 "name": "BaseBdev2", 00:16:16.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:16.305 "is_configured": false, 00:16:16.305 "data_offset": 0, 00:16:16.305 "data_size": 0 00:16:16.305 } 00:16:16.305 ] 00:16:16.305 }' 00:16:16.305 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.305 14:15:20 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.565 [2024-09-30 14:15:21.136294] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:16.565 [2024-09-30 14:15:21.136989] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:16:16.565 [2024-09-30 14:15:21.137067] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:16.565 [2024-09-30 14:15:21.137390] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:16.565 BaseBdev2 00:16:16.565 [2024-09-30 14:15:21.137639] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:16:16.565 [2024-09-30 14:15:21.137675] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:16:16.565 [2024-09-30 14:15:21.137854] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.565 [ 00:16:16.565 { 00:16:16.565 "name": "BaseBdev2", 00:16:16.565 "aliases": [ 00:16:16.565 "e3f64265-bbe9-4dc1-b854-3cc95d5106c1" 00:16:16.565 ], 00:16:16.565 "product_name": "Malloc disk", 00:16:16.565 "block_size": 4128, 00:16:16.565 "num_blocks": 8192, 00:16:16.565 "uuid": "e3f64265-bbe9-4dc1-b854-3cc95d5106c1", 00:16:16.565 "md_size": 32, 00:16:16.565 "md_interleave": true, 00:16:16.565 "dif_type": 0, 00:16:16.565 "assigned_rate_limits": { 00:16:16.565 "rw_ios_per_sec": 0, 00:16:16.565 "rw_mbytes_per_sec": 0, 00:16:16.565 "r_mbytes_per_sec": 0, 00:16:16.565 "w_mbytes_per_sec": 0 00:16:16.565 }, 00:16:16.565 "claimed": true, 00:16:16.565 "claim_type": "exclusive_write", 00:16:16.565 "zoned": false, 00:16:16.565 "supported_io_types": { 00:16:16.565 "read": true, 00:16:16.565 "write": true, 00:16:16.565 "unmap": true, 00:16:16.565 "flush": true, 00:16:16.565 "reset": true, 00:16:16.565 "nvme_admin": false, 00:16:16.565 "nvme_io": false, 00:16:16.565 "nvme_io_md": false, 00:16:16.565 "write_zeroes": true, 00:16:16.565 "zcopy": true, 00:16:16.565 "get_zone_info": false, 00:16:16.565 "zone_management": false, 00:16:16.565 "zone_append": false, 00:16:16.565 "compare": false, 00:16:16.565 "compare_and_write": false, 00:16:16.565 "abort": true, 00:16:16.565 "seek_hole": false, 00:16:16.565 "seek_data": false, 00:16:16.565 "copy": true, 00:16:16.565 "nvme_iov_md": false 00:16:16.565 }, 00:16:16.565 "memory_domains": [ 00:16:16.565 { 00:16:16.565 "dma_device_id": "system", 00:16:16.565 "dma_device_type": 1 00:16:16.565 }, 00:16:16.565 { 00:16:16.565 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:16.565 "dma_device_type": 2 00:16:16.565 } 00:16:16.565 ], 00:16:16.565 "driver_specific": {} 00:16:16.565 } 00:16:16.565 ] 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:16.565 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.566 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.825 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.825 "name": "Existed_Raid", 00:16:16.825 "uuid": "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1", 00:16:16.825 "strip_size_kb": 0, 00:16:16.825 "state": "online", 00:16:16.825 "raid_level": "raid1", 00:16:16.825 "superblock": true, 00:16:16.825 "num_base_bdevs": 2, 00:16:16.825 "num_base_bdevs_discovered": 2, 00:16:16.825 "num_base_bdevs_operational": 2, 00:16:16.825 "base_bdevs_list": [ 00:16:16.825 { 00:16:16.825 "name": "BaseBdev1", 00:16:16.825 "uuid": "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460", 00:16:16.825 "is_configured": true, 00:16:16.825 "data_offset": 256, 00:16:16.825 "data_size": 7936 00:16:16.825 }, 00:16:16.825 { 00:16:16.825 "name": "BaseBdev2", 00:16:16.825 "uuid": "e3f64265-bbe9-4dc1-b854-3cc95d5106c1", 00:16:16.825 "is_configured": true, 00:16:16.825 "data_offset": 256, 00:16:16.825 "data_size": 7936 00:16:16.825 } 00:16:16.825 ] 00:16:16.825 }' 00:16:16.825 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.825 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:17.085 [2024-09-30 14:15:21.631832] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:17.085 "name": "Existed_Raid", 00:16:17.085 "aliases": [ 00:16:17.085 "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1" 00:16:17.085 ], 00:16:17.085 "product_name": "Raid Volume", 00:16:17.085 "block_size": 4128, 00:16:17.085 "num_blocks": 7936, 00:16:17.085 "uuid": "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1", 00:16:17.085 "md_size": 32, 00:16:17.085 "md_interleave": true, 00:16:17.085 "dif_type": 0, 00:16:17.085 "assigned_rate_limits": { 00:16:17.085 "rw_ios_per_sec": 0, 00:16:17.085 "rw_mbytes_per_sec": 0, 00:16:17.085 "r_mbytes_per_sec": 0, 00:16:17.085 "w_mbytes_per_sec": 0 00:16:17.085 }, 00:16:17.085 "claimed": false, 00:16:17.085 "zoned": false, 00:16:17.085 "supported_io_types": { 00:16:17.085 "read": true, 00:16:17.085 "write": true, 00:16:17.085 "unmap": false, 00:16:17.085 "flush": false, 00:16:17.085 "reset": true, 00:16:17.085 "nvme_admin": false, 00:16:17.085 "nvme_io": false, 00:16:17.085 "nvme_io_md": false, 00:16:17.085 "write_zeroes": true, 00:16:17.085 "zcopy": false, 00:16:17.085 "get_zone_info": false, 00:16:17.085 "zone_management": false, 00:16:17.085 "zone_append": false, 00:16:17.085 "compare": false, 00:16:17.085 "compare_and_write": false, 00:16:17.085 "abort": false, 00:16:17.085 "seek_hole": false, 00:16:17.085 "seek_data": false, 00:16:17.085 "copy": false, 00:16:17.085 "nvme_iov_md": false 00:16:17.085 }, 00:16:17.085 "memory_domains": [ 00:16:17.085 { 00:16:17.085 "dma_device_id": "system", 00:16:17.085 "dma_device_type": 1 00:16:17.085 }, 00:16:17.085 { 00:16:17.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.085 "dma_device_type": 2 00:16:17.085 }, 00:16:17.085 { 00:16:17.085 "dma_device_id": "system", 00:16:17.085 "dma_device_type": 1 00:16:17.085 }, 00:16:17.085 { 00:16:17.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.085 "dma_device_type": 2 00:16:17.085 } 00:16:17.085 ], 00:16:17.085 "driver_specific": { 00:16:17.085 "raid": { 00:16:17.085 "uuid": "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1", 00:16:17.085 "strip_size_kb": 0, 00:16:17.085 "state": "online", 00:16:17.085 "raid_level": "raid1", 00:16:17.085 "superblock": true, 00:16:17.085 "num_base_bdevs": 2, 00:16:17.085 "num_base_bdevs_discovered": 2, 00:16:17.085 "num_base_bdevs_operational": 2, 00:16:17.085 "base_bdevs_list": [ 00:16:17.085 { 00:16:17.085 "name": "BaseBdev1", 00:16:17.085 "uuid": "2e499dc2-9be2-4c8b-ad0c-6fd6d36af460", 00:16:17.085 "is_configured": true, 00:16:17.085 "data_offset": 256, 00:16:17.085 "data_size": 7936 00:16:17.085 }, 00:16:17.085 { 00:16:17.085 "name": "BaseBdev2", 00:16:17.085 "uuid": "e3f64265-bbe9-4dc1-b854-3cc95d5106c1", 00:16:17.085 "is_configured": true, 00:16:17.085 "data_offset": 256, 00:16:17.085 "data_size": 7936 00:16:17.085 } 00:16:17.085 ] 00:16:17.085 } 00:16:17.085 } 00:16:17.085 }' 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:17.085 BaseBdev2' 00:16:17.085 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.346 [2024-09-30 14:15:21.855342] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.346 "name": "Existed_Raid", 00:16:17.346 "uuid": "aaaae8f1-d12c-4a4e-8c64-bbbcaf6518e1", 00:16:17.346 "strip_size_kb": 0, 00:16:17.346 "state": "online", 00:16:17.346 "raid_level": "raid1", 00:16:17.346 "superblock": true, 00:16:17.346 "num_base_bdevs": 2, 00:16:17.346 "num_base_bdevs_discovered": 1, 00:16:17.346 "num_base_bdevs_operational": 1, 00:16:17.346 "base_bdevs_list": [ 00:16:17.346 { 00:16:17.346 "name": null, 00:16:17.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:17.346 "is_configured": false, 00:16:17.346 "data_offset": 0, 00:16:17.346 "data_size": 7936 00:16:17.346 }, 00:16:17.346 { 00:16:17.346 "name": "BaseBdev2", 00:16:17.346 "uuid": "e3f64265-bbe9-4dc1-b854-3cc95d5106c1", 00:16:17.346 "is_configured": true, 00:16:17.346 "data_offset": 256, 00:16:17.346 "data_size": 7936 00:16:17.346 } 00:16:17.346 ] 00:16:17.346 }' 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.346 14:15:21 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.916 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.917 [2024-09-30 14:15:22.386240] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:17.917 [2024-09-30 14:15:22.386337] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:17.917 [2024-09-30 14:15:22.398379] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:17.917 [2024-09-30 14:15:22.398503] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:17.917 [2024-09-30 14:15:22.398544] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 99786 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 99786 ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 99786 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99786 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:17.917 killing process with pid 99786 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99786' 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 99786 00:16:17.917 [2024-09-30 14:15:22.487050] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:17.917 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 99786 00:16:17.917 [2024-09-30 14:15:22.488088] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:18.178 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:18.178 00:16:18.178 real 0m3.897s 00:16:18.178 user 0m6.046s 00:16:18.178 sys 0m0.872s 00:16:18.178 ************************************ 00:16:18.178 END TEST raid_state_function_test_sb_md_interleaved 00:16:18.178 ************************************ 00:16:18.178 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:18.178 14:15:22 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.178 14:15:22 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:18.178 14:15:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:16:18.178 14:15:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:18.178 14:15:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:18.178 ************************************ 00:16:18.178 START TEST raid_superblock_test_md_interleaved 00:16:18.178 ************************************ 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:18.178 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=100027 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 100027 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 100027 ']' 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:18.178 14:15:22 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.438 [2024-09-30 14:15:22.902411] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:18.439 [2024-09-30 14:15:22.902545] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100027 ] 00:16:18.439 [2024-09-30 14:15:23.032673] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:18.439 [2024-09-30 14:15:23.062475] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:18.698 [2024-09-30 14:15:23.110500] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:18.698 [2024-09-30 14:15:23.152980] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:18.698 [2024-09-30 14:15:23.153107] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.269 malloc1 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.269 [2024-09-30 14:15:23.735153] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:19.269 [2024-09-30 14:15:23.735278] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:19.269 [2024-09-30 14:15:23.735320] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:19.269 [2024-09-30 14:15:23.735348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:19.269 [2024-09-30 14:15:23.737335] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:19.269 [2024-09-30 14:15:23.737402] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:19.269 pt1 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:19.269 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.270 malloc2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.270 [2024-09-30 14:15:23.779068] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:19.270 [2024-09-30 14:15:23.779184] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:19.270 [2024-09-30 14:15:23.779229] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:19.270 [2024-09-30 14:15:23.779262] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:19.270 [2024-09-30 14:15:23.781651] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:19.270 [2024-09-30 14:15:23.781732] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:19.270 pt2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.270 [2024-09-30 14:15:23.791083] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:19.270 [2024-09-30 14:15:23.793013] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:19.270 [2024-09-30 14:15:23.793165] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:19.270 [2024-09-30 14:15:23.793178] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:19.270 [2024-09-30 14:15:23.793252] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:19.270 [2024-09-30 14:15:23.793317] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:19.270 [2024-09-30 14:15:23.793330] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:19.270 [2024-09-30 14:15:23.793397] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.270 "name": "raid_bdev1", 00:16:19.270 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:19.270 "strip_size_kb": 0, 00:16:19.270 "state": "online", 00:16:19.270 "raid_level": "raid1", 00:16:19.270 "superblock": true, 00:16:19.270 "num_base_bdevs": 2, 00:16:19.270 "num_base_bdevs_discovered": 2, 00:16:19.270 "num_base_bdevs_operational": 2, 00:16:19.270 "base_bdevs_list": [ 00:16:19.270 { 00:16:19.270 "name": "pt1", 00:16:19.270 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:19.270 "is_configured": true, 00:16:19.270 "data_offset": 256, 00:16:19.270 "data_size": 7936 00:16:19.270 }, 00:16:19.270 { 00:16:19.270 "name": "pt2", 00:16:19.270 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:19.270 "is_configured": true, 00:16:19.270 "data_offset": 256, 00:16:19.270 "data_size": 7936 00:16:19.270 } 00:16:19.270 ] 00:16:19.270 }' 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.270 14:15:23 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.840 [2024-09-30 14:15:24.274506] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:19.840 "name": "raid_bdev1", 00:16:19.840 "aliases": [ 00:16:19.840 "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16" 00:16:19.840 ], 00:16:19.840 "product_name": "Raid Volume", 00:16:19.840 "block_size": 4128, 00:16:19.840 "num_blocks": 7936, 00:16:19.840 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:19.840 "md_size": 32, 00:16:19.840 "md_interleave": true, 00:16:19.840 "dif_type": 0, 00:16:19.840 "assigned_rate_limits": { 00:16:19.840 "rw_ios_per_sec": 0, 00:16:19.840 "rw_mbytes_per_sec": 0, 00:16:19.840 "r_mbytes_per_sec": 0, 00:16:19.840 "w_mbytes_per_sec": 0 00:16:19.840 }, 00:16:19.840 "claimed": false, 00:16:19.840 "zoned": false, 00:16:19.840 "supported_io_types": { 00:16:19.840 "read": true, 00:16:19.840 "write": true, 00:16:19.840 "unmap": false, 00:16:19.840 "flush": false, 00:16:19.840 "reset": true, 00:16:19.840 "nvme_admin": false, 00:16:19.840 "nvme_io": false, 00:16:19.840 "nvme_io_md": false, 00:16:19.840 "write_zeroes": true, 00:16:19.840 "zcopy": false, 00:16:19.840 "get_zone_info": false, 00:16:19.840 "zone_management": false, 00:16:19.840 "zone_append": false, 00:16:19.840 "compare": false, 00:16:19.840 "compare_and_write": false, 00:16:19.840 "abort": false, 00:16:19.840 "seek_hole": false, 00:16:19.840 "seek_data": false, 00:16:19.840 "copy": false, 00:16:19.840 "nvme_iov_md": false 00:16:19.840 }, 00:16:19.840 "memory_domains": [ 00:16:19.840 { 00:16:19.840 "dma_device_id": "system", 00:16:19.840 "dma_device_type": 1 00:16:19.840 }, 00:16:19.840 { 00:16:19.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:19.840 "dma_device_type": 2 00:16:19.840 }, 00:16:19.840 { 00:16:19.840 "dma_device_id": "system", 00:16:19.840 "dma_device_type": 1 00:16:19.840 }, 00:16:19.840 { 00:16:19.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:19.840 "dma_device_type": 2 00:16:19.840 } 00:16:19.840 ], 00:16:19.840 "driver_specific": { 00:16:19.840 "raid": { 00:16:19.840 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:19.840 "strip_size_kb": 0, 00:16:19.840 "state": "online", 00:16:19.840 "raid_level": "raid1", 00:16:19.840 "superblock": true, 00:16:19.840 "num_base_bdevs": 2, 00:16:19.840 "num_base_bdevs_discovered": 2, 00:16:19.840 "num_base_bdevs_operational": 2, 00:16:19.840 "base_bdevs_list": [ 00:16:19.840 { 00:16:19.840 "name": "pt1", 00:16:19.840 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:19.840 "is_configured": true, 00:16:19.840 "data_offset": 256, 00:16:19.840 "data_size": 7936 00:16:19.840 }, 00:16:19.840 { 00:16:19.840 "name": "pt2", 00:16:19.840 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:19.840 "is_configured": true, 00:16:19.840 "data_offset": 256, 00:16:19.840 "data_size": 7936 00:16:19.840 } 00:16:19.840 ] 00:16:19.840 } 00:16:19.840 } 00:16:19.840 }' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:19.840 pt2' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.840 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 [2024-09-30 14:15:24.525983] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 ']' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 [2024-09-30 14:15:24.573672] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.101 [2024-09-30 14:15:24.573735] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:20.101 [2024-09-30 14:15:24.573827] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:20.101 [2024-09-30 14:15:24.573901] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:20.101 [2024-09-30 14:15:24.573933] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.101 [2024-09-30 14:15:24.717428] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:20.101 [2024-09-30 14:15:24.719191] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:20.101 [2024-09-30 14:15:24.719249] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:20.101 [2024-09-30 14:15:24.719296] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:20.101 [2024-09-30 14:15:24.719311] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.101 [2024-09-30 14:15:24.719320] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:16:20.101 request: 00:16:20.101 { 00:16:20.101 "name": "raid_bdev1", 00:16:20.101 "raid_level": "raid1", 00:16:20.101 "base_bdevs": [ 00:16:20.101 "malloc1", 00:16:20.101 "malloc2" 00:16:20.101 ], 00:16:20.101 "superblock": false, 00:16:20.101 "method": "bdev_raid_create", 00:16:20.101 "req_id": 1 00:16:20.101 } 00:16:20.101 Got JSON-RPC error response 00:16:20.101 response: 00:16:20.101 { 00:16:20.101 "code": -17, 00:16:20.101 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:20.101 } 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:20.101 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:20.102 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.102 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:20.102 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.102 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.102 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.362 [2024-09-30 14:15:24.781304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:20.362 [2024-09-30 14:15:24.781395] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.362 [2024-09-30 14:15:24.781424] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:20.362 [2024-09-30 14:15:24.781451] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.362 [2024-09-30 14:15:24.783200] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.362 [2024-09-30 14:15:24.783270] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:20.362 [2024-09-30 14:15:24.783329] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:20.362 [2024-09-30 14:15:24.783389] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:20.362 pt1 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.362 "name": "raid_bdev1", 00:16:20.362 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:20.362 "strip_size_kb": 0, 00:16:20.362 "state": "configuring", 00:16:20.362 "raid_level": "raid1", 00:16:20.362 "superblock": true, 00:16:20.362 "num_base_bdevs": 2, 00:16:20.362 "num_base_bdevs_discovered": 1, 00:16:20.362 "num_base_bdevs_operational": 2, 00:16:20.362 "base_bdevs_list": [ 00:16:20.362 { 00:16:20.362 "name": "pt1", 00:16:20.362 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:20.362 "is_configured": true, 00:16:20.362 "data_offset": 256, 00:16:20.362 "data_size": 7936 00:16:20.362 }, 00:16:20.362 { 00:16:20.362 "name": null, 00:16:20.362 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:20.362 "is_configured": false, 00:16:20.362 "data_offset": 256, 00:16:20.362 "data_size": 7936 00:16:20.362 } 00:16:20.362 ] 00:16:20.362 }' 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.362 14:15:24 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.622 [2024-09-30 14:15:25.228545] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:20.622 [2024-09-30 14:15:25.228642] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.622 [2024-09-30 14:15:25.228676] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:20.622 [2024-09-30 14:15:25.228705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.622 [2024-09-30 14:15:25.228851] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.622 [2024-09-30 14:15:25.228896] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:20.622 [2024-09-30 14:15:25.228960] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:20.622 [2024-09-30 14:15:25.229012] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:20.622 [2024-09-30 14:15:25.229104] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:16:20.622 [2024-09-30 14:15:25.229140] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:20.622 [2024-09-30 14:15:25.229217] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:16:20.622 [2024-09-30 14:15:25.229304] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:16:20.622 [2024-09-30 14:15:25.229334] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:16:20.622 [2024-09-30 14:15:25.229412] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.622 pt2 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.622 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.886 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.886 "name": "raid_bdev1", 00:16:20.886 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:20.886 "strip_size_kb": 0, 00:16:20.886 "state": "online", 00:16:20.886 "raid_level": "raid1", 00:16:20.886 "superblock": true, 00:16:20.886 "num_base_bdevs": 2, 00:16:20.886 "num_base_bdevs_discovered": 2, 00:16:20.886 "num_base_bdevs_operational": 2, 00:16:20.886 "base_bdevs_list": [ 00:16:20.886 { 00:16:20.886 "name": "pt1", 00:16:20.886 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:20.886 "is_configured": true, 00:16:20.886 "data_offset": 256, 00:16:20.886 "data_size": 7936 00:16:20.886 }, 00:16:20.886 { 00:16:20.886 "name": "pt2", 00:16:20.886 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:20.886 "is_configured": true, 00:16:20.886 "data_offset": 256, 00:16:20.886 "data_size": 7936 00:16:20.886 } 00:16:20.886 ] 00:16:20.886 }' 00:16:20.886 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.886 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.149 [2024-09-30 14:15:25.679980] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:21.149 "name": "raid_bdev1", 00:16:21.149 "aliases": [ 00:16:21.149 "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16" 00:16:21.149 ], 00:16:21.149 "product_name": "Raid Volume", 00:16:21.149 "block_size": 4128, 00:16:21.149 "num_blocks": 7936, 00:16:21.149 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:21.149 "md_size": 32, 00:16:21.149 "md_interleave": true, 00:16:21.149 "dif_type": 0, 00:16:21.149 "assigned_rate_limits": { 00:16:21.149 "rw_ios_per_sec": 0, 00:16:21.149 "rw_mbytes_per_sec": 0, 00:16:21.149 "r_mbytes_per_sec": 0, 00:16:21.149 "w_mbytes_per_sec": 0 00:16:21.149 }, 00:16:21.149 "claimed": false, 00:16:21.149 "zoned": false, 00:16:21.149 "supported_io_types": { 00:16:21.149 "read": true, 00:16:21.149 "write": true, 00:16:21.149 "unmap": false, 00:16:21.149 "flush": false, 00:16:21.149 "reset": true, 00:16:21.149 "nvme_admin": false, 00:16:21.149 "nvme_io": false, 00:16:21.149 "nvme_io_md": false, 00:16:21.149 "write_zeroes": true, 00:16:21.149 "zcopy": false, 00:16:21.149 "get_zone_info": false, 00:16:21.149 "zone_management": false, 00:16:21.149 "zone_append": false, 00:16:21.149 "compare": false, 00:16:21.149 "compare_and_write": false, 00:16:21.149 "abort": false, 00:16:21.149 "seek_hole": false, 00:16:21.149 "seek_data": false, 00:16:21.149 "copy": false, 00:16:21.149 "nvme_iov_md": false 00:16:21.149 }, 00:16:21.149 "memory_domains": [ 00:16:21.149 { 00:16:21.149 "dma_device_id": "system", 00:16:21.149 "dma_device_type": 1 00:16:21.149 }, 00:16:21.149 { 00:16:21.149 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:21.149 "dma_device_type": 2 00:16:21.149 }, 00:16:21.149 { 00:16:21.149 "dma_device_id": "system", 00:16:21.149 "dma_device_type": 1 00:16:21.149 }, 00:16:21.149 { 00:16:21.149 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:21.149 "dma_device_type": 2 00:16:21.149 } 00:16:21.149 ], 00:16:21.149 "driver_specific": { 00:16:21.149 "raid": { 00:16:21.149 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:21.149 "strip_size_kb": 0, 00:16:21.149 "state": "online", 00:16:21.149 "raid_level": "raid1", 00:16:21.149 "superblock": true, 00:16:21.149 "num_base_bdevs": 2, 00:16:21.149 "num_base_bdevs_discovered": 2, 00:16:21.149 "num_base_bdevs_operational": 2, 00:16:21.149 "base_bdevs_list": [ 00:16:21.149 { 00:16:21.149 "name": "pt1", 00:16:21.149 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:21.149 "is_configured": true, 00:16:21.149 "data_offset": 256, 00:16:21.149 "data_size": 7936 00:16:21.149 }, 00:16:21.149 { 00:16:21.149 "name": "pt2", 00:16:21.149 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:21.149 "is_configured": true, 00:16:21.149 "data_offset": 256, 00:16:21.149 "data_size": 7936 00:16:21.149 } 00:16:21.149 ] 00:16:21.149 } 00:16:21.149 } 00:16:21.149 }' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:21.149 pt2' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.149 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.442 [2024-09-30 14:15:25.907780] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 '!=' 3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 ']' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.442 [2024-09-30 14:15:25.939627] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:21.442 "name": "raid_bdev1", 00:16:21.442 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:21.442 "strip_size_kb": 0, 00:16:21.442 "state": "online", 00:16:21.442 "raid_level": "raid1", 00:16:21.442 "superblock": true, 00:16:21.442 "num_base_bdevs": 2, 00:16:21.442 "num_base_bdevs_discovered": 1, 00:16:21.442 "num_base_bdevs_operational": 1, 00:16:21.442 "base_bdevs_list": [ 00:16:21.442 { 00:16:21.442 "name": null, 00:16:21.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:21.442 "is_configured": false, 00:16:21.442 "data_offset": 0, 00:16:21.442 "data_size": 7936 00:16:21.442 }, 00:16:21.442 { 00:16:21.442 "name": "pt2", 00:16:21.442 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:21.442 "is_configured": true, 00:16:21.442 "data_offset": 256, 00:16:21.442 "data_size": 7936 00:16:21.442 } 00:16:21.442 ] 00:16:21.442 }' 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:21.442 14:15:25 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.709 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:21.709 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.709 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.709 [2024-09-30 14:15:26.355571] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:21.709 [2024-09-30 14:15:26.355639] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:21.709 [2024-09-30 14:15:26.355710] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:21.709 [2024-09-30 14:15:26.355761] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:21.709 [2024-09-30 14:15:26.355795] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:16:21.709 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 [2024-09-30 14:15:26.423594] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:21.972 [2024-09-30 14:15:26.423683] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:21.972 [2024-09-30 14:15:26.423711] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:21.972 [2024-09-30 14:15:26.423739] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:21.972 [2024-09-30 14:15:26.425613] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:21.972 [2024-09-30 14:15:26.425680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:21.972 [2024-09-30 14:15:26.425743] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:21.972 [2024-09-30 14:15:26.425790] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:21.972 [2024-09-30 14:15:26.425866] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:16:21.972 [2024-09-30 14:15:26.425908] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:21.972 [2024-09-30 14:15:26.426021] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:21.972 [2024-09-30 14:15:26.426115] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:16:21.972 [2024-09-30 14:15:26.426148] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:16:21.972 [2024-09-30 14:15:26.426239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:21.972 pt2 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:21.972 "name": "raid_bdev1", 00:16:21.972 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:21.972 "strip_size_kb": 0, 00:16:21.972 "state": "online", 00:16:21.972 "raid_level": "raid1", 00:16:21.972 "superblock": true, 00:16:21.972 "num_base_bdevs": 2, 00:16:21.972 "num_base_bdevs_discovered": 1, 00:16:21.972 "num_base_bdevs_operational": 1, 00:16:21.972 "base_bdevs_list": [ 00:16:21.972 { 00:16:21.972 "name": null, 00:16:21.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:21.972 "is_configured": false, 00:16:21.972 "data_offset": 256, 00:16:21.972 "data_size": 7936 00:16:21.972 }, 00:16:21.972 { 00:16:21.972 "name": "pt2", 00:16:21.972 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:21.972 "is_configured": true, 00:16:21.972 "data_offset": 256, 00:16:21.972 "data_size": 7936 00:16:21.972 } 00:16:21.972 ] 00:16:21.972 }' 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:21.972 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.232 [2024-09-30 14:15:26.875574] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:22.232 [2024-09-30 14:15:26.875646] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:22.232 [2024-09-30 14:15:26.875695] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:22.232 [2024-09-30 14:15:26.875732] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:22.232 [2024-09-30 14:15:26.875740] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.232 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.491 [2024-09-30 14:15:26.939601] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:22.491 [2024-09-30 14:15:26.939646] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:22.491 [2024-09-30 14:15:26.939665] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:22.491 [2024-09-30 14:15:26.939673] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:22.491 [2024-09-30 14:15:26.941446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:22.491 [2024-09-30 14:15:26.941493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:22.491 [2024-09-30 14:15:26.941533] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:22.491 [2024-09-30 14:15:26.941565] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:22.491 [2024-09-30 14:15:26.941647] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:22.491 [2024-09-30 14:15:26.941658] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:22.491 [2024-09-30 14:15:26.941673] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:16:22.491 [2024-09-30 14:15:26.941696] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:22.491 [2024-09-30 14:15:26.941745] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:16:22.491 [2024-09-30 14:15:26.941752] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:22.491 [2024-09-30 14:15:26.941810] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:16:22.491 [2024-09-30 14:15:26.941869] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:16:22.491 [2024-09-30 14:15:26.941879] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:16:22.491 [2024-09-30 14:15:26.941935] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:22.491 pt1 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:22.491 "name": "raid_bdev1", 00:16:22.491 "uuid": "3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16", 00:16:22.491 "strip_size_kb": 0, 00:16:22.491 "state": "online", 00:16:22.491 "raid_level": "raid1", 00:16:22.491 "superblock": true, 00:16:22.491 "num_base_bdevs": 2, 00:16:22.491 "num_base_bdevs_discovered": 1, 00:16:22.491 "num_base_bdevs_operational": 1, 00:16:22.491 "base_bdevs_list": [ 00:16:22.491 { 00:16:22.491 "name": null, 00:16:22.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.491 "is_configured": false, 00:16:22.491 "data_offset": 256, 00:16:22.491 "data_size": 7936 00:16:22.491 }, 00:16:22.491 { 00:16:22.491 "name": "pt2", 00:16:22.491 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:22.491 "is_configured": true, 00:16:22.491 "data_offset": 256, 00:16:22.491 "data_size": 7936 00:16:22.491 } 00:16:22.491 ] 00:16:22.491 }' 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:22.491 14:15:26 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.060 [2024-09-30 14:15:27.487772] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 '!=' 3c2e9b9f-0616-4a20-9a94-f4a7ecdc4b16 ']' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 100027 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 100027 ']' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 100027 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100027 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100027' 00:16:23.060 killing process with pid 100027 00:16:23.060 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@969 -- # kill 100027 00:16:23.060 [2024-09-30 14:15:27.554812] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:23.060 [2024-09-30 14:15:27.554921] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:23.060 [2024-09-30 14:15:27.554981] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:23.060 [2024-09-30 14:15:27.555029] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, sta 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@974 -- # wait 100027 00:16:23.060 te offline 00:16:23.060 [2024-09-30 14:15:27.577885] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:23.319 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:16:23.319 00:16:23.319 real 0m5.002s 00:16:23.319 user 0m8.186s 00:16:23.319 sys 0m1.086s 00:16:23.319 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:23.320 ************************************ 00:16:23.320 END TEST raid_superblock_test_md_interleaved 00:16:23.320 ************************************ 00:16:23.320 14:15:27 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.320 14:15:27 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:16:23.320 14:15:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:23.320 14:15:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:23.320 14:15:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:23.320 ************************************ 00:16:23.320 START TEST raid_rebuild_test_sb_md_interleaved 00:16:23.320 ************************************ 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false false 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=100340 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 100340 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 100340 ']' 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:23.320 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:23.320 14:15:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.579 [2024-09-30 14:15:28.000137] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:23.580 [2024-09-30 14:15:28.000356] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100340 ] 00:16:23.580 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:23.580 Zero copy mechanism will not be used. 00:16:23.580 [2024-09-30 14:15:28.136946] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:23.580 [2024-09-30 14:15:28.164378] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:23.580 [2024-09-30 14:15:28.210026] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:23.839 [2024-09-30 14:15:28.252511] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:23.839 [2024-09-30 14:15:28.252546] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.408 BaseBdev1_malloc 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.408 [2024-09-30 14:15:28.810934] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:24.408 [2024-09-30 14:15:28.811023] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:24.408 [2024-09-30 14:15:28.811051] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:24.408 [2024-09-30 14:15:28.811064] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:24.408 [2024-09-30 14:15:28.812932] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:24.408 [2024-09-30 14:15:28.813054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:24.408 BaseBdev1 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.408 BaseBdev2_malloc 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.408 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.408 [2024-09-30 14:15:28.844073] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:24.408 [2024-09-30 14:15:28.844128] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:24.408 [2024-09-30 14:15:28.844146] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:24.408 [2024-09-30 14:15:28.844157] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:24.408 [2024-09-30 14:15:28.845980] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:24.408 [2024-09-30 14:15:28.846021] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:24.408 BaseBdev2 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.409 spare_malloc 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.409 spare_delay 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.409 [2024-09-30 14:15:28.884673] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:24.409 [2024-09-30 14:15:28.884728] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:24.409 [2024-09-30 14:15:28.884747] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:24.409 [2024-09-30 14:15:28.884757] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:24.409 [2024-09-30 14:15:28.886566] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:24.409 [2024-09-30 14:15:28.886682] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:24.409 spare 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.409 [2024-09-30 14:15:28.896707] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:24.409 [2024-09-30 14:15:28.898486] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:24.409 [2024-09-30 14:15:28.898649] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:24.409 [2024-09-30 14:15:28.898664] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:24.409 [2024-09-30 14:15:28.898730] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:16:24.409 [2024-09-30 14:15:28.898796] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:24.409 [2024-09-30 14:15:28.898805] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:24.409 [2024-09-30 14:15:28.898864] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.409 "name": "raid_bdev1", 00:16:24.409 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:24.409 "strip_size_kb": 0, 00:16:24.409 "state": "online", 00:16:24.409 "raid_level": "raid1", 00:16:24.409 "superblock": true, 00:16:24.409 "num_base_bdevs": 2, 00:16:24.409 "num_base_bdevs_discovered": 2, 00:16:24.409 "num_base_bdevs_operational": 2, 00:16:24.409 "base_bdevs_list": [ 00:16:24.409 { 00:16:24.409 "name": "BaseBdev1", 00:16:24.409 "uuid": "adde376d-9c64-53f2-8338-c1c1825fe063", 00:16:24.409 "is_configured": true, 00:16:24.409 "data_offset": 256, 00:16:24.409 "data_size": 7936 00:16:24.409 }, 00:16:24.409 { 00:16:24.409 "name": "BaseBdev2", 00:16:24.409 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:24.409 "is_configured": true, 00:16:24.409 "data_offset": 256, 00:16:24.409 "data_size": 7936 00:16:24.409 } 00:16:24.409 ] 00:16:24.409 }' 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.409 14:15:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.976 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.977 [2024-09-30 14:15:29.368147] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.977 [2024-09-30 14:15:29.463710] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.977 "name": "raid_bdev1", 00:16:24.977 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:24.977 "strip_size_kb": 0, 00:16:24.977 "state": "online", 00:16:24.977 "raid_level": "raid1", 00:16:24.977 "superblock": true, 00:16:24.977 "num_base_bdevs": 2, 00:16:24.977 "num_base_bdevs_discovered": 1, 00:16:24.977 "num_base_bdevs_operational": 1, 00:16:24.977 "base_bdevs_list": [ 00:16:24.977 { 00:16:24.977 "name": null, 00:16:24.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.977 "is_configured": false, 00:16:24.977 "data_offset": 0, 00:16:24.977 "data_size": 7936 00:16:24.977 }, 00:16:24.977 { 00:16:24.977 "name": "BaseBdev2", 00:16:24.977 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:24.977 "is_configured": true, 00:16:24.977 "data_offset": 256, 00:16:24.977 "data_size": 7936 00:16:24.977 } 00:16:24.977 ] 00:16:24.977 }' 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.977 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.546 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:25.546 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.546 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.546 [2024-09-30 14:15:29.923608] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:25.546 [2024-09-30 14:15:29.926472] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:25.546 [2024-09-30 14:15:29.928290] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:25.546 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.546 14:15:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.484 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.484 "name": "raid_bdev1", 00:16:26.484 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:26.484 "strip_size_kb": 0, 00:16:26.484 "state": "online", 00:16:26.484 "raid_level": "raid1", 00:16:26.484 "superblock": true, 00:16:26.484 "num_base_bdevs": 2, 00:16:26.484 "num_base_bdevs_discovered": 2, 00:16:26.484 "num_base_bdevs_operational": 2, 00:16:26.484 "process": { 00:16:26.484 "type": "rebuild", 00:16:26.484 "target": "spare", 00:16:26.484 "progress": { 00:16:26.484 "blocks": 2560, 00:16:26.484 "percent": 32 00:16:26.484 } 00:16:26.484 }, 00:16:26.484 "base_bdevs_list": [ 00:16:26.484 { 00:16:26.485 "name": "spare", 00:16:26.485 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:26.485 "is_configured": true, 00:16:26.485 "data_offset": 256, 00:16:26.485 "data_size": 7936 00:16:26.485 }, 00:16:26.485 { 00:16:26.485 "name": "BaseBdev2", 00:16:26.485 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:26.485 "is_configured": true, 00:16:26.485 "data_offset": 256, 00:16:26.485 "data_size": 7936 00:16:26.485 } 00:16:26.485 ] 00:16:26.485 }' 00:16:26.485 14:15:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.485 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.485 [2024-09-30 14:15:31.091648] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.485 [2024-09-30 14:15:31.132926] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:26.485 [2024-09-30 14:15:31.133045] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:26.485 [2024-09-30 14:15:31.133079] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.485 [2024-09-30 14:15:31.133103] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.744 "name": "raid_bdev1", 00:16:26.744 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:26.744 "strip_size_kb": 0, 00:16:26.744 "state": "online", 00:16:26.744 "raid_level": "raid1", 00:16:26.744 "superblock": true, 00:16:26.744 "num_base_bdevs": 2, 00:16:26.744 "num_base_bdevs_discovered": 1, 00:16:26.744 "num_base_bdevs_operational": 1, 00:16:26.744 "base_bdevs_list": [ 00:16:26.744 { 00:16:26.744 "name": null, 00:16:26.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.744 "is_configured": false, 00:16:26.744 "data_offset": 0, 00:16:26.744 "data_size": 7936 00:16:26.744 }, 00:16:26.744 { 00:16:26.744 "name": "BaseBdev2", 00:16:26.744 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:26.744 "is_configured": true, 00:16:26.744 "data_offset": 256, 00:16:26.744 "data_size": 7936 00:16:26.744 } 00:16:26.744 ] 00:16:26.744 }' 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.744 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.004 "name": "raid_bdev1", 00:16:27.004 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:27.004 "strip_size_kb": 0, 00:16:27.004 "state": "online", 00:16:27.004 "raid_level": "raid1", 00:16:27.004 "superblock": true, 00:16:27.004 "num_base_bdevs": 2, 00:16:27.004 "num_base_bdevs_discovered": 1, 00:16:27.004 "num_base_bdevs_operational": 1, 00:16:27.004 "base_bdevs_list": [ 00:16:27.004 { 00:16:27.004 "name": null, 00:16:27.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.004 "is_configured": false, 00:16:27.004 "data_offset": 0, 00:16:27.004 "data_size": 7936 00:16:27.004 }, 00:16:27.004 { 00:16:27.004 "name": "BaseBdev2", 00:16:27.004 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:27.004 "is_configured": true, 00:16:27.004 "data_offset": 256, 00:16:27.004 "data_size": 7936 00:16:27.004 } 00:16:27.004 ] 00:16:27.004 }' 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:27.004 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.264 [2024-09-30 14:15:31.687616] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:27.264 [2024-09-30 14:15:31.690145] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:16:27.264 [2024-09-30 14:15:31.691952] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.264 14:15:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.201 "name": "raid_bdev1", 00:16:28.201 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:28.201 "strip_size_kb": 0, 00:16:28.201 "state": "online", 00:16:28.201 "raid_level": "raid1", 00:16:28.201 "superblock": true, 00:16:28.201 "num_base_bdevs": 2, 00:16:28.201 "num_base_bdevs_discovered": 2, 00:16:28.201 "num_base_bdevs_operational": 2, 00:16:28.201 "process": { 00:16:28.201 "type": "rebuild", 00:16:28.201 "target": "spare", 00:16:28.201 "progress": { 00:16:28.201 "blocks": 2560, 00:16:28.201 "percent": 32 00:16:28.201 } 00:16:28.201 }, 00:16:28.201 "base_bdevs_list": [ 00:16:28.201 { 00:16:28.201 "name": "spare", 00:16:28.201 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:28.201 "is_configured": true, 00:16:28.201 "data_offset": 256, 00:16:28.201 "data_size": 7936 00:16:28.201 }, 00:16:28.201 { 00:16:28.201 "name": "BaseBdev2", 00:16:28.201 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:28.201 "is_configured": true, 00:16:28.201 "data_offset": 256, 00:16:28.201 "data_size": 7936 00:16:28.201 } 00:16:28.201 ] 00:16:28.201 }' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:28.201 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=613 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:28.201 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.461 "name": "raid_bdev1", 00:16:28.461 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:28.461 "strip_size_kb": 0, 00:16:28.461 "state": "online", 00:16:28.461 "raid_level": "raid1", 00:16:28.461 "superblock": true, 00:16:28.461 "num_base_bdevs": 2, 00:16:28.461 "num_base_bdevs_discovered": 2, 00:16:28.461 "num_base_bdevs_operational": 2, 00:16:28.461 "process": { 00:16:28.461 "type": "rebuild", 00:16:28.461 "target": "spare", 00:16:28.461 "progress": { 00:16:28.461 "blocks": 2816, 00:16:28.461 "percent": 35 00:16:28.461 } 00:16:28.461 }, 00:16:28.461 "base_bdevs_list": [ 00:16:28.461 { 00:16:28.461 "name": "spare", 00:16:28.461 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:28.461 "is_configured": true, 00:16:28.461 "data_offset": 256, 00:16:28.461 "data_size": 7936 00:16:28.461 }, 00:16:28.461 { 00:16:28.461 "name": "BaseBdev2", 00:16:28.461 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:28.461 "is_configured": true, 00:16:28.461 "data_offset": 256, 00:16:28.461 "data_size": 7936 00:16:28.461 } 00:16:28.461 ] 00:16:28.461 }' 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:28.461 14:15:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.399 14:15:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.399 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.399 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:29.399 "name": "raid_bdev1", 00:16:29.399 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:29.399 "strip_size_kb": 0, 00:16:29.399 "state": "online", 00:16:29.399 "raid_level": "raid1", 00:16:29.399 "superblock": true, 00:16:29.399 "num_base_bdevs": 2, 00:16:29.399 "num_base_bdevs_discovered": 2, 00:16:29.399 "num_base_bdevs_operational": 2, 00:16:29.399 "process": { 00:16:29.399 "type": "rebuild", 00:16:29.399 "target": "spare", 00:16:29.399 "progress": { 00:16:29.399 "blocks": 5888, 00:16:29.399 "percent": 74 00:16:29.399 } 00:16:29.399 }, 00:16:29.399 "base_bdevs_list": [ 00:16:29.399 { 00:16:29.399 "name": "spare", 00:16:29.399 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:29.399 "is_configured": true, 00:16:29.399 "data_offset": 256, 00:16:29.399 "data_size": 7936 00:16:29.399 }, 00:16:29.399 { 00:16:29.399 "name": "BaseBdev2", 00:16:29.399 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:29.399 "is_configured": true, 00:16:29.399 "data_offset": 256, 00:16:29.399 "data_size": 7936 00:16:29.399 } 00:16:29.399 ] 00:16:29.399 }' 00:16:29.399 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:29.658 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:29.658 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:29.658 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:29.658 14:15:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:30.227 [2024-09-30 14:15:34.802173] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:30.227 [2024-09-30 14:15:34.802244] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:30.227 [2024-09-30 14:15:34.802333] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.485 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:30.744 "name": "raid_bdev1", 00:16:30.744 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:30.744 "strip_size_kb": 0, 00:16:30.744 "state": "online", 00:16:30.744 "raid_level": "raid1", 00:16:30.744 "superblock": true, 00:16:30.744 "num_base_bdevs": 2, 00:16:30.744 "num_base_bdevs_discovered": 2, 00:16:30.744 "num_base_bdevs_operational": 2, 00:16:30.744 "base_bdevs_list": [ 00:16:30.744 { 00:16:30.744 "name": "spare", 00:16:30.744 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:30.744 "is_configured": true, 00:16:30.744 "data_offset": 256, 00:16:30.744 "data_size": 7936 00:16:30.744 }, 00:16:30.744 { 00:16:30.744 "name": "BaseBdev2", 00:16:30.744 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:30.744 "is_configured": true, 00:16:30.744 "data_offset": 256, 00:16:30.744 "data_size": 7936 00:16:30.744 } 00:16:30.744 ] 00:16:30.744 }' 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.744 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:30.744 "name": "raid_bdev1", 00:16:30.744 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:30.744 "strip_size_kb": 0, 00:16:30.744 "state": "online", 00:16:30.744 "raid_level": "raid1", 00:16:30.744 "superblock": true, 00:16:30.744 "num_base_bdevs": 2, 00:16:30.744 "num_base_bdevs_discovered": 2, 00:16:30.744 "num_base_bdevs_operational": 2, 00:16:30.744 "base_bdevs_list": [ 00:16:30.744 { 00:16:30.744 "name": "spare", 00:16:30.744 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:30.745 "is_configured": true, 00:16:30.745 "data_offset": 256, 00:16:30.745 "data_size": 7936 00:16:30.745 }, 00:16:30.745 { 00:16:30.745 "name": "BaseBdev2", 00:16:30.745 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:30.745 "is_configured": true, 00:16:30.745 "data_offset": 256, 00:16:30.745 "data_size": 7936 00:16:30.745 } 00:16:30.745 ] 00:16:30.745 }' 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.745 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.004 "name": "raid_bdev1", 00:16:31.004 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:31.004 "strip_size_kb": 0, 00:16:31.004 "state": "online", 00:16:31.004 "raid_level": "raid1", 00:16:31.004 "superblock": true, 00:16:31.004 "num_base_bdevs": 2, 00:16:31.004 "num_base_bdevs_discovered": 2, 00:16:31.004 "num_base_bdevs_operational": 2, 00:16:31.004 "base_bdevs_list": [ 00:16:31.004 { 00:16:31.004 "name": "spare", 00:16:31.004 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:31.004 "is_configured": true, 00:16:31.004 "data_offset": 256, 00:16:31.004 "data_size": 7936 00:16:31.004 }, 00:16:31.004 { 00:16:31.004 "name": "BaseBdev2", 00:16:31.004 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:31.004 "is_configured": true, 00:16:31.004 "data_offset": 256, 00:16:31.004 "data_size": 7936 00:16:31.004 } 00:16:31.004 ] 00:16:31.004 }' 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.004 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.263 [2024-09-30 14:15:35.835432] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:31.263 [2024-09-30 14:15:35.835563] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:31.263 [2024-09-30 14:15:35.835650] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:31.263 [2024-09-30 14:15:35.835715] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:31.263 [2024-09-30 14:15:35.835725] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.263 [2024-09-30 14:15:35.903317] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:31.263 [2024-09-30 14:15:35.903378] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:31.263 [2024-09-30 14:15:35.903398] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:31.263 [2024-09-30 14:15:35.903407] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:31.263 [2024-09-30 14:15:35.905368] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:31.263 [2024-09-30 14:15:35.905408] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:31.263 [2024-09-30 14:15:35.905456] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:31.263 [2024-09-30 14:15:35.905500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:31.263 [2024-09-30 14:15:35.905596] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:31.263 spare 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.263 14:15:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.522 [2024-09-30 14:15:36.005496] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:16:31.522 [2024-09-30 14:15:36.005525] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:31.522 [2024-09-30 14:15:36.005610] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:16:31.522 [2024-09-30 14:15:36.005680] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:16:31.522 [2024-09-30 14:15:36.005690] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:16:31.522 [2024-09-30 14:15:36.005756] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.522 "name": "raid_bdev1", 00:16:31.522 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:31.522 "strip_size_kb": 0, 00:16:31.522 "state": "online", 00:16:31.522 "raid_level": "raid1", 00:16:31.522 "superblock": true, 00:16:31.522 "num_base_bdevs": 2, 00:16:31.522 "num_base_bdevs_discovered": 2, 00:16:31.522 "num_base_bdevs_operational": 2, 00:16:31.522 "base_bdevs_list": [ 00:16:31.522 { 00:16:31.522 "name": "spare", 00:16:31.522 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:31.522 "is_configured": true, 00:16:31.522 "data_offset": 256, 00:16:31.522 "data_size": 7936 00:16:31.522 }, 00:16:31.522 { 00:16:31.522 "name": "BaseBdev2", 00:16:31.522 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:31.522 "is_configured": true, 00:16:31.522 "data_offset": 256, 00:16:31.522 "data_size": 7936 00:16:31.522 } 00:16:31.522 ] 00:16:31.522 }' 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.522 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:32.091 "name": "raid_bdev1", 00:16:32.091 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:32.091 "strip_size_kb": 0, 00:16:32.091 "state": "online", 00:16:32.091 "raid_level": "raid1", 00:16:32.091 "superblock": true, 00:16:32.091 "num_base_bdevs": 2, 00:16:32.091 "num_base_bdevs_discovered": 2, 00:16:32.091 "num_base_bdevs_operational": 2, 00:16:32.091 "base_bdevs_list": [ 00:16:32.091 { 00:16:32.091 "name": "spare", 00:16:32.091 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:32.091 "is_configured": true, 00:16:32.091 "data_offset": 256, 00:16:32.091 "data_size": 7936 00:16:32.091 }, 00:16:32.091 { 00:16:32.091 "name": "BaseBdev2", 00:16:32.091 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:32.091 "is_configured": true, 00:16:32.091 "data_offset": 256, 00:16:32.091 "data_size": 7936 00:16:32.091 } 00:16:32.091 ] 00:16:32.091 }' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.091 [2024-09-30 14:15:36.638077] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.091 "name": "raid_bdev1", 00:16:32.091 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:32.091 "strip_size_kb": 0, 00:16:32.091 "state": "online", 00:16:32.091 "raid_level": "raid1", 00:16:32.091 "superblock": true, 00:16:32.091 "num_base_bdevs": 2, 00:16:32.091 "num_base_bdevs_discovered": 1, 00:16:32.091 "num_base_bdevs_operational": 1, 00:16:32.091 "base_bdevs_list": [ 00:16:32.091 { 00:16:32.091 "name": null, 00:16:32.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.091 "is_configured": false, 00:16:32.091 "data_offset": 0, 00:16:32.091 "data_size": 7936 00:16:32.091 }, 00:16:32.091 { 00:16:32.091 "name": "BaseBdev2", 00:16:32.091 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:32.091 "is_configured": true, 00:16:32.091 "data_offset": 256, 00:16:32.091 "data_size": 7936 00:16:32.091 } 00:16:32.091 ] 00:16:32.091 }' 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.091 14:15:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.660 14:15:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:32.660 14:15:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.660 14:15:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.660 [2024-09-30 14:15:37.141239] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:32.660 [2024-09-30 14:15:37.141409] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:32.660 [2024-09-30 14:15:37.141468] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:32.660 [2024-09-30 14:15:37.141549] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:32.660 [2024-09-30 14:15:37.144283] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:16:32.660 [2024-09-30 14:15:37.146052] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:32.660 14:15:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.660 14:15:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:33.599 "name": "raid_bdev1", 00:16:33.599 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:33.599 "strip_size_kb": 0, 00:16:33.599 "state": "online", 00:16:33.599 "raid_level": "raid1", 00:16:33.599 "superblock": true, 00:16:33.599 "num_base_bdevs": 2, 00:16:33.599 "num_base_bdevs_discovered": 2, 00:16:33.599 "num_base_bdevs_operational": 2, 00:16:33.599 "process": { 00:16:33.599 "type": "rebuild", 00:16:33.599 "target": "spare", 00:16:33.599 "progress": { 00:16:33.599 "blocks": 2560, 00:16:33.599 "percent": 32 00:16:33.599 } 00:16:33.599 }, 00:16:33.599 "base_bdevs_list": [ 00:16:33.599 { 00:16:33.599 "name": "spare", 00:16:33.599 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:33.599 "is_configured": true, 00:16:33.599 "data_offset": 256, 00:16:33.599 "data_size": 7936 00:16:33.599 }, 00:16:33.599 { 00:16:33.599 "name": "BaseBdev2", 00:16:33.599 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:33.599 "is_configured": true, 00:16:33.599 "data_offset": 256, 00:16:33.599 "data_size": 7936 00:16:33.599 } 00:16:33.599 ] 00:16:33.599 }' 00:16:33.599 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.860 [2024-09-30 14:15:38.292789] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:33.860 [2024-09-30 14:15:38.350051] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:33.860 [2024-09-30 14:15:38.350107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:33.860 [2024-09-30 14:15:38.350120] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:33.860 [2024-09-30 14:15:38.350128] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:33.860 "name": "raid_bdev1", 00:16:33.860 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:33.860 "strip_size_kb": 0, 00:16:33.860 "state": "online", 00:16:33.860 "raid_level": "raid1", 00:16:33.860 "superblock": true, 00:16:33.860 "num_base_bdevs": 2, 00:16:33.860 "num_base_bdevs_discovered": 1, 00:16:33.860 "num_base_bdevs_operational": 1, 00:16:33.860 "base_bdevs_list": [ 00:16:33.860 { 00:16:33.860 "name": null, 00:16:33.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.860 "is_configured": false, 00:16:33.860 "data_offset": 0, 00:16:33.860 "data_size": 7936 00:16:33.860 }, 00:16:33.860 { 00:16:33.860 "name": "BaseBdev2", 00:16:33.860 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:33.860 "is_configured": true, 00:16:33.860 "data_offset": 256, 00:16:33.860 "data_size": 7936 00:16:33.860 } 00:16:33.860 ] 00:16:33.860 }' 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:33.860 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.120 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:34.120 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.120 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.380 [2024-09-30 14:15:38.776566] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:34.380 [2024-09-30 14:15:38.776666] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.380 [2024-09-30 14:15:38.776706] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:16:34.380 [2024-09-30 14:15:38.776736] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.380 [2024-09-30 14:15:38.776939] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.380 [2024-09-30 14:15:38.776990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:34.380 [2024-09-30 14:15:38.777061] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:34.380 [2024-09-30 14:15:38.777099] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:34.380 [2024-09-30 14:15:38.777145] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:34.380 [2024-09-30 14:15:38.777211] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:34.380 [2024-09-30 14:15:38.779633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:16:34.380 [2024-09-30 14:15:38.781482] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:34.380 spare 00:16:34.380 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.380 14:15:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:35.317 "name": "raid_bdev1", 00:16:35.317 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:35.317 "strip_size_kb": 0, 00:16:35.317 "state": "online", 00:16:35.317 "raid_level": "raid1", 00:16:35.317 "superblock": true, 00:16:35.317 "num_base_bdevs": 2, 00:16:35.317 "num_base_bdevs_discovered": 2, 00:16:35.317 "num_base_bdevs_operational": 2, 00:16:35.317 "process": { 00:16:35.317 "type": "rebuild", 00:16:35.317 "target": "spare", 00:16:35.317 "progress": { 00:16:35.317 "blocks": 2560, 00:16:35.317 "percent": 32 00:16:35.317 } 00:16:35.317 }, 00:16:35.317 "base_bdevs_list": [ 00:16:35.317 { 00:16:35.317 "name": "spare", 00:16:35.317 "uuid": "02fdd245-e7a6-5bab-885c-380088597903", 00:16:35.317 "is_configured": true, 00:16:35.317 "data_offset": 256, 00:16:35.317 "data_size": 7936 00:16:35.317 }, 00:16:35.317 { 00:16:35.317 "name": "BaseBdev2", 00:16:35.317 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:35.317 "is_configured": true, 00:16:35.317 "data_offset": 256, 00:16:35.317 "data_size": 7936 00:16:35.317 } 00:16:35.317 ] 00:16:35.317 }' 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.317 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.317 [2024-09-30 14:15:39.932104] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:35.577 [2024-09-30 14:15:39.985309] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:35.577 [2024-09-30 14:15:39.985362] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:35.577 [2024-09-30 14:15:39.985378] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:35.577 [2024-09-30 14:15:39.985385] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.577 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.577 14:15:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.577 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.577 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.577 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:35.577 "name": "raid_bdev1", 00:16:35.577 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:35.577 "strip_size_kb": 0, 00:16:35.577 "state": "online", 00:16:35.577 "raid_level": "raid1", 00:16:35.578 "superblock": true, 00:16:35.578 "num_base_bdevs": 2, 00:16:35.578 "num_base_bdevs_discovered": 1, 00:16:35.578 "num_base_bdevs_operational": 1, 00:16:35.578 "base_bdevs_list": [ 00:16:35.578 { 00:16:35.578 "name": null, 00:16:35.578 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:35.578 "is_configured": false, 00:16:35.578 "data_offset": 0, 00:16:35.578 "data_size": 7936 00:16:35.578 }, 00:16:35.578 { 00:16:35.578 "name": "BaseBdev2", 00:16:35.578 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:35.578 "is_configured": true, 00:16:35.578 "data_offset": 256, 00:16:35.578 "data_size": 7936 00:16:35.578 } 00:16:35.578 ] 00:16:35.578 }' 00:16:35.578 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:35.578 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.837 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:36.097 "name": "raid_bdev1", 00:16:36.097 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:36.097 "strip_size_kb": 0, 00:16:36.097 "state": "online", 00:16:36.097 "raid_level": "raid1", 00:16:36.097 "superblock": true, 00:16:36.097 "num_base_bdevs": 2, 00:16:36.097 "num_base_bdevs_discovered": 1, 00:16:36.097 "num_base_bdevs_operational": 1, 00:16:36.097 "base_bdevs_list": [ 00:16:36.097 { 00:16:36.097 "name": null, 00:16:36.097 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.097 "is_configured": false, 00:16:36.097 "data_offset": 0, 00:16:36.097 "data_size": 7936 00:16:36.097 }, 00:16:36.097 { 00:16:36.097 "name": "BaseBdev2", 00:16:36.097 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:36.097 "is_configured": true, 00:16:36.097 "data_offset": 256, 00:16:36.097 "data_size": 7936 00:16:36.097 } 00:16:36.097 ] 00:16:36.097 }' 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.097 [2024-09-30 14:15:40.619474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:36.097 [2024-09-30 14:15:40.619561] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.097 [2024-09-30 14:15:40.619582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:16:36.097 [2024-09-30 14:15:40.619591] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.097 [2024-09-30 14:15:40.619725] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.097 [2024-09-30 14:15:40.619737] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:36.097 [2024-09-30 14:15:40.619779] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:36.097 [2024-09-30 14:15:40.619791] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:36.097 [2024-09-30 14:15:40.619805] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:36.097 [2024-09-30 14:15:40.619814] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:36.097 BaseBdev1 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.097 14:15:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:37.037 "name": "raid_bdev1", 00:16:37.037 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:37.037 "strip_size_kb": 0, 00:16:37.037 "state": "online", 00:16:37.037 "raid_level": "raid1", 00:16:37.037 "superblock": true, 00:16:37.037 "num_base_bdevs": 2, 00:16:37.037 "num_base_bdevs_discovered": 1, 00:16:37.037 "num_base_bdevs_operational": 1, 00:16:37.037 "base_bdevs_list": [ 00:16:37.037 { 00:16:37.037 "name": null, 00:16:37.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.037 "is_configured": false, 00:16:37.037 "data_offset": 0, 00:16:37.037 "data_size": 7936 00:16:37.037 }, 00:16:37.037 { 00:16:37.037 "name": "BaseBdev2", 00:16:37.037 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:37.037 "is_configured": true, 00:16:37.037 "data_offset": 256, 00:16:37.037 "data_size": 7936 00:16:37.037 } 00:16:37.037 ] 00:16:37.037 }' 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:37.037 14:15:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.605 "name": "raid_bdev1", 00:16:37.605 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:37.605 "strip_size_kb": 0, 00:16:37.605 "state": "online", 00:16:37.605 "raid_level": "raid1", 00:16:37.605 "superblock": true, 00:16:37.605 "num_base_bdevs": 2, 00:16:37.605 "num_base_bdevs_discovered": 1, 00:16:37.605 "num_base_bdevs_operational": 1, 00:16:37.605 "base_bdevs_list": [ 00:16:37.605 { 00:16:37.605 "name": null, 00:16:37.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.605 "is_configured": false, 00:16:37.605 "data_offset": 0, 00:16:37.605 "data_size": 7936 00:16:37.605 }, 00:16:37.605 { 00:16:37.605 "name": "BaseBdev2", 00:16:37.605 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:37.605 "is_configured": true, 00:16:37.605 "data_offset": 256, 00:16:37.605 "data_size": 7936 00:16:37.605 } 00:16:37.605 ] 00:16:37.605 }' 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:37.605 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.606 [2024-09-30 14:15:42.236728] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:37.606 [2024-09-30 14:15:42.236842] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:37.606 [2024-09-30 14:15:42.236863] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:37.606 request: 00:16:37.606 { 00:16:37.606 "base_bdev": "BaseBdev1", 00:16:37.606 "raid_bdev": "raid_bdev1", 00:16:37.606 "method": "bdev_raid_add_base_bdev", 00:16:37.606 "req_id": 1 00:16:37.606 } 00:16:37.606 Got JSON-RPC error response 00:16:37.606 response: 00:16:37.606 { 00:16:37.606 "code": -22, 00:16:37.606 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:37.606 } 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:37.606 14:15:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:38.983 "name": "raid_bdev1", 00:16:38.983 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:38.983 "strip_size_kb": 0, 00:16:38.983 "state": "online", 00:16:38.983 "raid_level": "raid1", 00:16:38.983 "superblock": true, 00:16:38.983 "num_base_bdevs": 2, 00:16:38.983 "num_base_bdevs_discovered": 1, 00:16:38.983 "num_base_bdevs_operational": 1, 00:16:38.983 "base_bdevs_list": [ 00:16:38.983 { 00:16:38.983 "name": null, 00:16:38.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.983 "is_configured": false, 00:16:38.983 "data_offset": 0, 00:16:38.983 "data_size": 7936 00:16:38.983 }, 00:16:38.983 { 00:16:38.983 "name": "BaseBdev2", 00:16:38.983 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:38.983 "is_configured": true, 00:16:38.983 "data_offset": 256, 00:16:38.983 "data_size": 7936 00:16:38.983 } 00:16:38.983 ] 00:16:38.983 }' 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:38.983 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.241 "name": "raid_bdev1", 00:16:39.241 "uuid": "581bead9-b117-431c-bca9-d36549ecd890", 00:16:39.241 "strip_size_kb": 0, 00:16:39.241 "state": "online", 00:16:39.241 "raid_level": "raid1", 00:16:39.241 "superblock": true, 00:16:39.241 "num_base_bdevs": 2, 00:16:39.241 "num_base_bdevs_discovered": 1, 00:16:39.241 "num_base_bdevs_operational": 1, 00:16:39.241 "base_bdevs_list": [ 00:16:39.241 { 00:16:39.241 "name": null, 00:16:39.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.241 "is_configured": false, 00:16:39.241 "data_offset": 0, 00:16:39.241 "data_size": 7936 00:16:39.241 }, 00:16:39.241 { 00:16:39.241 "name": "BaseBdev2", 00:16:39.241 "uuid": "57359236-177f-5913-98fe-c089b59529ee", 00:16:39.241 "is_configured": true, 00:16:39.241 "data_offset": 256, 00:16:39.241 "data_size": 7936 00:16:39.241 } 00:16:39.241 ] 00:16:39.241 }' 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:39.241 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 100340 00:16:39.242 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 100340 ']' 00:16:39.242 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 100340 00:16:39.242 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:39.242 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:39.242 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100340 00:16:39.501 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:39.501 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:39.501 killing process with pid 100340 00:16:39.501 Received shutdown signal, test time was about 60.000000 seconds 00:16:39.501 00:16:39.501 Latency(us) 00:16:39.501 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:39.501 =================================================================================================================== 00:16:39.501 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:39.501 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100340' 00:16:39.501 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 100340 00:16:39.501 [2024-09-30 14:15:43.905185] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:39.501 [2024-09-30 14:15:43.905291] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:39.501 [2024-09-30 14:15:43.905328] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:39.501 [2024-09-30 14:15:43.905339] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:16:39.501 14:15:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 100340 00:16:39.501 [2024-09-30 14:15:43.938358] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:39.760 14:15:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:16:39.760 00:16:39.760 real 0m16.276s 00:16:39.760 user 0m21.832s 00:16:39.760 sys 0m1.684s 00:16:39.760 14:15:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:39.760 14:15:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.760 ************************************ 00:16:39.761 END TEST raid_rebuild_test_sb_md_interleaved 00:16:39.761 ************************************ 00:16:39.761 14:15:44 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:16:39.761 14:15:44 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:16:39.761 14:15:44 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 100340 ']' 00:16:39.761 14:15:44 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 100340 00:16:39.761 14:15:44 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:16:39.761 00:16:39.761 real 9m54.329s 00:16:39.761 user 14m2.170s 00:16:39.761 sys 1m49.689s 00:16:39.761 ************************************ 00:16:39.761 END TEST bdev_raid 00:16:39.761 ************************************ 00:16:39.761 14:15:44 bdev_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:39.761 14:15:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:39.761 14:15:44 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:39.761 14:15:44 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:39.761 14:15:44 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:39.761 14:15:44 -- common/autotest_common.sh@10 -- # set +x 00:16:39.761 ************************************ 00:16:39.761 START TEST spdkcli_raid 00:16:39.761 ************************************ 00:16:39.761 14:15:44 spdkcli_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:40.021 * Looking for test storage... 00:16:40.021 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:40.021 14:15:44 spdkcli_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:40.021 14:15:44 spdkcli_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:16:40.021 14:15:44 spdkcli_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:40.021 14:15:44 spdkcli_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:40.021 14:15:44 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:40.022 14:15:44 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:40.022 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:40.022 --rc genhtml_branch_coverage=1 00:16:40.022 --rc genhtml_function_coverage=1 00:16:40.022 --rc genhtml_legend=1 00:16:40.022 --rc geninfo_all_blocks=1 00:16:40.022 --rc geninfo_unexecuted_blocks=1 00:16:40.022 00:16:40.022 ' 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:40.022 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:40.022 --rc genhtml_branch_coverage=1 00:16:40.022 --rc genhtml_function_coverage=1 00:16:40.022 --rc genhtml_legend=1 00:16:40.022 --rc geninfo_all_blocks=1 00:16:40.022 --rc geninfo_unexecuted_blocks=1 00:16:40.022 00:16:40.022 ' 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:40.022 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:40.022 --rc genhtml_branch_coverage=1 00:16:40.022 --rc genhtml_function_coverage=1 00:16:40.022 --rc genhtml_legend=1 00:16:40.022 --rc geninfo_all_blocks=1 00:16:40.022 --rc geninfo_unexecuted_blocks=1 00:16:40.022 00:16:40.022 ' 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:40.022 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:40.022 --rc genhtml_branch_coverage=1 00:16:40.022 --rc genhtml_function_coverage=1 00:16:40.022 --rc genhtml_legend=1 00:16:40.022 --rc geninfo_all_blocks=1 00:16:40.022 --rc geninfo_unexecuted_blocks=1 00:16:40.022 00:16:40.022 ' 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:16:40.022 14:15:44 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=101010 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:16:40.022 14:15:44 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 101010 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@831 -- # '[' -z 101010 ']' 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:40.022 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:40.022 14:15:44 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:40.282 [2024-09-30 14:15:44.713868] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:40.282 [2024-09-30 14:15:44.714092] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101010 ] 00:16:40.282 [2024-09-30 14:15:44.851016] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:40.282 [2024-09-30 14:15:44.881412] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:40.282 [2024-09-30 14:15:44.930872] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:40.282 [2024-09-30 14:15:44.930957] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@864 -- # return 0 00:16:41.222 14:15:45 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:41.222 14:15:45 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:41.222 14:15:45 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:41.222 14:15:45 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:16:41.222 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:16:41.222 ' 00:16:42.603 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:16:42.603 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:16:42.603 14:15:47 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:16:42.603 14:15:47 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:42.603 14:15:47 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.603 14:15:47 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:16:42.603 14:15:47 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:42.603 14:15:47 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.603 14:15:47 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:16:42.603 ' 00:16:43.993 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:16:43.993 14:15:48 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:16:43.993 14:15:48 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:43.993 14:15:48 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:43.993 14:15:48 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:16:43.993 14:15:48 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:43.993 14:15:48 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:43.993 14:15:48 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:16:43.993 14:15:48 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:16:44.562 14:15:48 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:16:44.562 14:15:48 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:16:44.562 14:15:48 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:16:44.562 14:15:48 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:44.562 14:15:48 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:44.562 14:15:49 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:16:44.562 14:15:49 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:44.562 14:15:49 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:44.562 14:15:49 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:16:44.562 ' 00:16:45.500 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:16:45.500 14:15:50 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:16:45.500 14:15:50 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:45.500 14:15:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:45.759 14:15:50 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:16:45.759 14:15:50 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:45.759 14:15:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:45.759 14:15:50 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:16:45.759 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:16:45.759 ' 00:16:47.140 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:16:47.140 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:16:47.140 14:15:51 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:47.140 14:15:51 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 101010 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 101010 ']' 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 101010 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@955 -- # uname 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 101010 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 101010' 00:16:47.140 killing process with pid 101010 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@969 -- # kill 101010 00:16:47.140 14:15:51 spdkcli_raid -- common/autotest_common.sh@974 -- # wait 101010 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 101010 ']' 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 101010 00:16:47.711 14:15:52 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 101010 ']' 00:16:47.711 14:15:52 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 101010 00:16:47.711 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (101010) - No such process 00:16:47.711 14:15:52 spdkcli_raid -- common/autotest_common.sh@977 -- # echo 'Process with pid 101010 is not found' 00:16:47.711 Process with pid 101010 is not found 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:16:47.711 14:15:52 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:16:47.711 00:16:47.711 real 0m7.750s 00:16:47.711 user 0m16.246s 00:16:47.711 sys 0m1.132s 00:16:47.711 14:15:52 spdkcli_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:47.711 14:15:52 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:47.711 ************************************ 00:16:47.711 END TEST spdkcli_raid 00:16:47.711 ************************************ 00:16:47.711 14:15:52 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:47.711 14:15:52 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:47.711 14:15:52 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:47.711 14:15:52 -- common/autotest_common.sh@10 -- # set +x 00:16:47.711 ************************************ 00:16:47.711 START TEST blockdev_raid5f 00:16:47.711 ************************************ 00:16:47.711 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:47.711 * Looking for test storage... 00:16:47.711 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:16:47.711 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:47.711 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lcov --version 00:16:47.711 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:47.971 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:16:47.971 14:15:52 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:47.972 14:15:52 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:47.972 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.972 --rc genhtml_branch_coverage=1 00:16:47.972 --rc genhtml_function_coverage=1 00:16:47.972 --rc genhtml_legend=1 00:16:47.972 --rc geninfo_all_blocks=1 00:16:47.972 --rc geninfo_unexecuted_blocks=1 00:16:47.972 00:16:47.972 ' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:47.972 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.972 --rc genhtml_branch_coverage=1 00:16:47.972 --rc genhtml_function_coverage=1 00:16:47.972 --rc genhtml_legend=1 00:16:47.972 --rc geninfo_all_blocks=1 00:16:47.972 --rc geninfo_unexecuted_blocks=1 00:16:47.972 00:16:47.972 ' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:47.972 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.972 --rc genhtml_branch_coverage=1 00:16:47.972 --rc genhtml_function_coverage=1 00:16:47.972 --rc genhtml_legend=1 00:16:47.972 --rc geninfo_all_blocks=1 00:16:47.972 --rc geninfo_unexecuted_blocks=1 00:16:47.972 00:16:47.972 ' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:47.972 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:47.972 --rc genhtml_branch_coverage=1 00:16:47.972 --rc genhtml_function_coverage=1 00:16:47.972 --rc genhtml_legend=1 00:16:47.972 --rc geninfo_all_blocks=1 00:16:47.972 --rc geninfo_unexecuted_blocks=1 00:16:47.972 00:16:47.972 ' 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@669 -- # QOS_DEV_1=Malloc_0 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@670 -- # QOS_DEV_2=Null_1 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_RUN_TIME=5 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@673 -- # uname -s 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@673 -- # '[' Linux = Linux ']' 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@675 -- # PRE_RESERVED_MEM=0 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@681 -- # test_type=raid5f 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@682 -- # crypto_device= 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@683 -- # dek= 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@684 -- # env_ctx= 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@685 -- # wait_for_rpc= 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@686 -- # '[' -n '' ']' 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == bdev ]] 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == crypto_* ]] 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@692 -- # start_spdk_tgt 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=101272 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 101272 00:16:47.972 14:15:52 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@831 -- # '[' -z 101272 ']' 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:47.972 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:47.972 14:15:52 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:47.972 [2024-09-30 14:15:52.526002] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:47.972 [2024-09-30 14:15:52.526149] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101272 ] 00:16:48.232 [2024-09-30 14:15:52.662989] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:48.232 [2024-09-30 14:15:52.689864] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:48.232 [2024-09-30 14:15:52.737686] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@864 -- # return 0 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@693 -- # case "$test_type" in 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@725 -- # setup_raid5f_conf 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:48.802 Malloc0 00:16:48.802 Malloc1 00:16:48.802 Malloc2 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@736 -- # rpc_cmd bdev_wait_for_examine 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@739 -- # cat 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n accel 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n bdev 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.802 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n iobuf 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.802 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@747 -- # mapfile -t bdevs 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@747 -- # rpc_cmd bdev_get_bdevs 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@747 -- # jq -r '.[] | select(.claimed == false)' 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@748 -- # mapfile -t bdevs_name 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@748 -- # jq -r .name 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@748 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "6db32858-ef8f-433c-a855-f7bf5bde002f"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "6db32858-ef8f-433c-a855-f7bf5bde002f",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "6db32858-ef8f-433c-a855-f7bf5bde002f",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "c8459fe1-a37c-4436-811e-333c947d7bda",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "3f7df9d4-29d0-423e-be54-8a0630ae6c7d",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "982b5d34-e56c-4202-867b-e4178c0427b8",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@749 -- # bdev_list=("${bdevs_name[@]}") 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@751 -- # hello_world_bdev=raid5f 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@752 -- # trap - SIGINT SIGTERM EXIT 00:16:49.062 14:15:53 blockdev_raid5f -- bdev/blockdev.sh@753 -- # killprocess 101272 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@950 -- # '[' -z 101272 ']' 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@954 -- # kill -0 101272 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@955 -- # uname 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 101272 00:16:49.062 killing process with pid 101272 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@968 -- # echo 'killing process with pid 101272' 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@969 -- # kill 101272 00:16:49.062 14:15:53 blockdev_raid5f -- common/autotest_common.sh@974 -- # wait 101272 00:16:49.632 14:15:54 blockdev_raid5f -- bdev/blockdev.sh@757 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:49.632 14:15:54 blockdev_raid5f -- bdev/blockdev.sh@759 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:49.632 14:15:54 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:49.632 14:15:54 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:49.632 14:15:54 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:49.632 ************************************ 00:16:49.632 START TEST bdev_hello_world 00:16:49.632 ************************************ 00:16:49.632 14:15:54 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:49.632 [2024-09-30 14:15:54.108805] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:49.632 [2024-09-30 14:15:54.108946] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101308 ] 00:16:49.632 [2024-09-30 14:15:54.244875] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:49.632 [2024-09-30 14:15:54.276013] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:49.893 [2024-09-30 14:15:54.324194] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:49.893 [2024-09-30 14:15:54.520686] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:16:49.893 [2024-09-30 14:15:54.520739] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:16:49.893 [2024-09-30 14:15:54.520766] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:16:49.893 [2024-09-30 14:15:54.521116] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:16:49.893 [2024-09-30 14:15:54.521257] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:16:49.893 [2024-09-30 14:15:54.521275] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:16:49.893 [2024-09-30 14:15:54.521327] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:16:49.893 00:16:49.893 [2024-09-30 14:15:54.521346] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:16:50.153 00:16:50.153 real 0m0.755s 00:16:50.153 user 0m0.402s 00:16:50.153 sys 0m0.237s 00:16:50.153 ************************************ 00:16:50.153 END TEST bdev_hello_world 00:16:50.153 ************************************ 00:16:50.153 14:15:54 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:50.153 14:15:54 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:16:50.413 14:15:54 blockdev_raid5f -- bdev/blockdev.sh@760 -- # run_test bdev_bounds bdev_bounds '' 00:16:50.413 14:15:54 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:50.413 14:15:54 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:50.413 14:15:54 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:50.413 ************************************ 00:16:50.413 START TEST bdev_bounds 00:16:50.413 ************************************ 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1125 -- # bdev_bounds '' 00:16:50.413 Process bdevio pid: 101339 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=101339 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 101339' 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 101339 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@831 -- # '[' -z 101339 ']' 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:50.413 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:50.413 14:15:54 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:50.413 [2024-09-30 14:15:54.942971] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:50.413 [2024-09-30 14:15:54.943220] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101339 ] 00:16:50.672 [2024-09-30 14:15:55.080687] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:50.672 [2024-09-30 14:15:55.108299] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:16:50.672 [2024-09-30 14:15:55.156908] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:16:50.672 [2024-09-30 14:15:55.157011] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:50.672 [2024-09-30 14:15:55.157124] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:16:51.241 14:15:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:51.241 14:15:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # return 0 00:16:51.241 14:15:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:16:51.241 I/O targets: 00:16:51.241 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:16:51.241 00:16:51.241 00:16:51.241 CUnit - A unit testing framework for C - Version 2.1-3 00:16:51.241 http://cunit.sourceforge.net/ 00:16:51.241 00:16:51.241 00:16:51.241 Suite: bdevio tests on: raid5f 00:16:51.241 Test: blockdev write read block ...passed 00:16:51.241 Test: blockdev write zeroes read block ...passed 00:16:51.241 Test: blockdev write zeroes read no split ...passed 00:16:51.500 Test: blockdev write zeroes read split ...passed 00:16:51.500 Test: blockdev write zeroes read split partial ...passed 00:16:51.501 Test: blockdev reset ...passed 00:16:51.501 Test: blockdev write read 8 blocks ...passed 00:16:51.501 Test: blockdev write read size > 128k ...passed 00:16:51.501 Test: blockdev write read invalid size ...passed 00:16:51.501 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:16:51.501 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:16:51.501 Test: blockdev write read max offset ...passed 00:16:51.501 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:16:51.501 Test: blockdev writev readv 8 blocks ...passed 00:16:51.501 Test: blockdev writev readv 30 x 1block ...passed 00:16:51.501 Test: blockdev writev readv block ...passed 00:16:51.501 Test: blockdev writev readv size > 128k ...passed 00:16:51.501 Test: blockdev writev readv size > 128k in two iovs ...passed 00:16:51.501 Test: blockdev comparev and writev ...passed 00:16:51.501 Test: blockdev nvme passthru rw ...passed 00:16:51.501 Test: blockdev nvme passthru vendor specific ...passed 00:16:51.501 Test: blockdev nvme admin passthru ...passed 00:16:51.501 Test: blockdev copy ...passed 00:16:51.501 00:16:51.501 Run Summary: Type Total Ran Passed Failed Inactive 00:16:51.501 suites 1 1 n/a 0 0 00:16:51.501 tests 23 23 23 0 0 00:16:51.501 asserts 130 130 130 0 n/a 00:16:51.501 00:16:51.501 Elapsed time = 0.310 seconds 00:16:51.501 0 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 101339 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@950 -- # '[' -z 101339 ']' 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # kill -0 101339 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # uname 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 101339 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@968 -- # echo 'killing process with pid 101339' 00:16:51.501 killing process with pid 101339 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@969 -- # kill 101339 00:16:51.501 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@974 -- # wait 101339 00:16:51.761 14:15:56 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:16:51.761 00:16:51.761 real 0m1.486s 00:16:51.761 user 0m3.501s 00:16:51.761 sys 0m0.382s 00:16:51.761 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:51.761 14:15:56 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:51.761 ************************************ 00:16:51.761 END TEST bdev_bounds 00:16:51.761 ************************************ 00:16:51.761 14:15:56 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:51.761 14:15:56 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:51.761 14:15:56 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:51.761 14:15:56 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:51.761 ************************************ 00:16:51.761 START TEST bdev_nbd 00:16:51.761 ************************************ 00:16:51.761 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1125 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=101387 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 101387 /var/tmp/spdk-nbd.sock 00:16:52.026 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@831 -- # '[' -z 101387 ']' 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:52.026 14:15:56 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:52.026 [2024-09-30 14:15:56.511452] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:16:52.026 [2024-09-30 14:15:56.511606] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:52.026 [2024-09-30 14:15:56.648125] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:16:52.026 [2024-09-30 14:15:56.675216] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:52.295 [2024-09-30 14:15:56.724936] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # return 0 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:52.888 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:53.157 1+0 records in 00:16:53.157 1+0 records out 00:16:53.157 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000407468 s, 10.1 MB/s 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:16:53.157 { 00:16:53.157 "nbd_device": "/dev/nbd0", 00:16:53.157 "bdev_name": "raid5f" 00:16:53.157 } 00:16:53.157 ]' 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:16:53.157 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:16:53.157 { 00:16:53.157 "nbd_device": "/dev/nbd0", 00:16:53.157 "bdev_name": "raid5f" 00:16:53.157 } 00:16:53.157 ]' 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:53.416 14:15:57 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:53.416 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:53.676 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:53.676 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:53.676 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:53.676 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:53.676 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:16:53.935 /dev/nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:16:53.935 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:54.195 1+0 records in 00:16:54.195 1+0 records out 00:16:54.195 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000411623 s, 10.0 MB/s 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:16:54.195 { 00:16:54.195 "nbd_device": "/dev/nbd0", 00:16:54.195 "bdev_name": "raid5f" 00:16:54.195 } 00:16:54.195 ]' 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:54.195 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:16:54.195 { 00:16:54.195 "nbd_device": "/dev/nbd0", 00:16:54.195 "bdev_name": "raid5f" 00:16:54.195 } 00:16:54.195 ]' 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:54.454 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:16:54.455 256+0 records in 00:16:54.455 256+0 records out 00:16:54.455 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0111814 s, 93.8 MB/s 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:16:54.455 256+0 records in 00:16:54.455 256+0 records out 00:16:54.455 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0266785 s, 39.3 MB/s 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:54.455 14:15:58 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:54.714 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:16:54.973 malloc_lvol_verify 00:16:54.973 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:16:55.232 bec601e1-f8c5-44d9-93ea-a95e1c62f600 00:16:55.232 14:15:59 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:16:55.491 df7ffdbb-e6da-440b-bd7e-89bc3f1aeff6 00:16:55.491 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:16:55.751 /dev/nbd0 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:16:55.751 mke2fs 1.47.0 (5-Feb-2023) 00:16:55.751 Discarding device blocks: 0/4096 done 00:16:55.751 Creating filesystem with 4096 1k blocks and 1024 inodes 00:16:55.751 00:16:55.751 Allocating group tables: 0/1 done 00:16:55.751 Writing inode tables: 0/1 done 00:16:55.751 Creating journal (1024 blocks): done 00:16:55.751 Writing superblocks and filesystem accounting information: 0/1 done 00:16:55.751 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:55.751 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 101387 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@950 -- # '[' -z 101387 ']' 00:16:56.010 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # kill -0 101387 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # uname 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 101387 00:16:56.011 killing process with pid 101387 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@968 -- # echo 'killing process with pid 101387' 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@969 -- # kill 101387 00:16:56.011 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@974 -- # wait 101387 00:16:56.271 ************************************ 00:16:56.271 END TEST bdev_nbd 00:16:56.271 ************************************ 00:16:56.271 14:16:00 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:16:56.271 00:16:56.271 real 0m4.408s 00:16:56.271 user 0m6.396s 00:16:56.271 sys 0m1.282s 00:16:56.271 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:56.271 14:16:00 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:56.271 14:16:00 blockdev_raid5f -- bdev/blockdev.sh@762 -- # [[ y == y ]] 00:16:56.271 14:16:00 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = nvme ']' 00:16:56.271 14:16:00 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = gpt ']' 00:16:56.271 14:16:00 blockdev_raid5f -- bdev/blockdev.sh@767 -- # run_test bdev_fio fio_test_suite '' 00:16:56.271 14:16:00 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:56.271 14:16:00 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:56.271 14:16:00 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:56.271 ************************************ 00:16:56.271 START TEST bdev_fio 00:16:56.271 ************************************ 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1125 -- # fio_test_suite '' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:16:56.271 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:16:56.271 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:16:56.531 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:16:56.531 14:16:00 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1101 -- # '[' 11 -le 1 ']' 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:56.531 ************************************ 00:16:56.531 START TEST bdev_fio_rw_verify 00:16:56.531 ************************************ 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1125 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib=/usr/lib64/libasan.so.8 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # break 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:16:56.531 14:16:01 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:56.791 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:16:56.791 fio-3.35 00:16:56.791 Starting 1 thread 00:17:09.069 00:17:09.069 job_raid5f: (groupid=0, jobs=1): err= 0: pid=101584: Mon Sep 30 14:16:11 2024 00:17:09.069 read: IOPS=12.3k, BW=48.2MiB/s (50.5MB/s)(482MiB/10001msec) 00:17:09.069 slat (usec): min=16, max=177, avg=19.05, stdev= 2.83 00:17:09.069 clat (usec): min=11, max=1012, avg=129.79, stdev=47.81 00:17:09.069 lat (usec): min=31, max=1190, avg=148.84, stdev=48.71 00:17:09.069 clat percentiles (usec): 00:17:09.069 | 50.000th=[ 135], 99.000th=[ 215], 99.900th=[ 367], 99.990th=[ 832], 00:17:09.069 | 99.999th=[ 963] 00:17:09.069 write: IOPS=12.9k, BW=50.6MiB/s (53.0MB/s)(499MiB/9875msec); 0 zone resets 00:17:09.069 slat (usec): min=8, max=311, avg=16.76, stdev= 4.11 00:17:09.069 clat (usec): min=56, max=1750, avg=297.53, stdev=43.12 00:17:09.069 lat (usec): min=72, max=2023, avg=314.29, stdev=44.29 00:17:09.069 clat percentiles (usec): 00:17:09.069 | 50.000th=[ 302], 99.000th=[ 379], 99.900th=[ 611], 99.990th=[ 1369], 00:17:09.069 | 99.999th=[ 1663] 00:17:09.069 bw ( KiB/s): min=48464, max=55024, per=98.66%, avg=51074.53, stdev=1569.14, samples=19 00:17:09.069 iops : min=12116, max=13756, avg=12768.63, stdev=392.29, samples=19 00:17:09.069 lat (usec) : 20=0.01%, 50=0.01%, 100=16.93%, 250=38.91%, 500=44.06% 00:17:09.069 lat (usec) : 750=0.06%, 1000=0.03% 00:17:09.069 lat (msec) : 2=0.01% 00:17:09.069 cpu : usr=98.68%, sys=0.54%, ctx=31, majf=0, minf=13196 00:17:09.069 IO depths : 1=7.6%, 2=19.8%, 4=55.3%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:09.069 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:09.069 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:09.069 issued rwts: total=123403,127806,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:09.069 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:09.069 00:17:09.069 Run status group 0 (all jobs): 00:17:09.069 READ: bw=48.2MiB/s (50.5MB/s), 48.2MiB/s-48.2MiB/s (50.5MB/s-50.5MB/s), io=482MiB (505MB), run=10001-10001msec 00:17:09.069 WRITE: bw=50.6MiB/s (53.0MB/s), 50.6MiB/s-50.6MiB/s (53.0MB/s-53.0MB/s), io=499MiB (523MB), run=9875-9875msec 00:17:09.069 ----------------------------------------------------- 00:17:09.069 Suppressions used: 00:17:09.069 count bytes template 00:17:09.069 1 7 /usr/src/fio/parse.c 00:17:09.069 574 55104 /usr/src/fio/iolog.c 00:17:09.069 1 8 libtcmalloc_minimal.so 00:17:09.069 1 904 libcrypto.so 00:17:09.069 ----------------------------------------------------- 00:17:09.069 00:17:09.069 00:17:09.069 real 0m11.236s 00:17:09.069 user 0m11.505s 00:17:09.069 sys 0m0.483s 00:17:09.069 14:16:12 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:09.069 14:16:12 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:09.069 ************************************ 00:17:09.069 END TEST bdev_fio_rw_verify 00:17:09.070 ************************************ 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "6db32858-ef8f-433c-a855-f7bf5bde002f"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "6db32858-ef8f-433c-a855-f7bf5bde002f",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "6db32858-ef8f-433c-a855-f7bf5bde002f",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "c8459fe1-a37c-4436-811e-333c947d7bda",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "3f7df9d4-29d0-423e-be54-8a0630ae6c7d",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "982b5d34-e56c-4202-867b-e4178c0427b8",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:09.070 /home/vagrant/spdk_repo/spdk 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:09.070 00:17:09.070 real 0m11.524s 00:17:09.070 user 0m11.630s 00:17:09.070 sys 0m0.620s 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:09.070 14:16:12 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:09.070 ************************************ 00:17:09.070 END TEST bdev_fio 00:17:09.070 ************************************ 00:17:09.070 14:16:12 blockdev_raid5f -- bdev/blockdev.sh@774 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:09.070 14:16:12 blockdev_raid5f -- bdev/blockdev.sh@776 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:09.070 14:16:12 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:09.070 14:16:12 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:09.070 14:16:12 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:09.070 ************************************ 00:17:09.070 START TEST bdev_verify 00:17:09.070 ************************************ 00:17:09.070 14:16:12 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:09.070 [2024-09-30 14:16:12.584119] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:17:09.070 [2024-09-30 14:16:12.584232] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101736 ] 00:17:09.070 [2024-09-30 14:16:12.717199] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:17:09.070 [2024-09-30 14:16:12.747275] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:09.070 [2024-09-30 14:16:12.803125] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:09.070 [2024-09-30 14:16:12.803230] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:09.070 Running I/O for 5 seconds... 00:17:13.393 14121.00 IOPS, 55.16 MiB/s 12815.50 IOPS, 50.06 MiB/s 12278.00 IOPS, 47.96 MiB/s 12059.75 IOPS, 47.11 MiB/s 11909.40 IOPS, 46.52 MiB/s 00:17:13.393 Latency(us) 00:17:13.393 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:13.393 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:13.393 Verification LBA range: start 0x0 length 0x2000 00:17:13.393 raid5f : 5.03 4971.84 19.42 0.00 0.00 38673.46 250.41 30449.91 00:17:13.393 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:13.393 Verification LBA range: start 0x2000 length 0x2000 00:17:13.393 raid5f : 5.02 6930.41 27.07 0.00 0.00 27717.04 122.52 36860.42 00:17:13.393 =================================================================================================================== 00:17:13.393 Total : 11902.25 46.49 0.00 0.00 32295.30 122.52 36860.42 00:17:13.653 00:17:13.653 real 0m5.777s 00:17:13.653 user 0m10.700s 00:17:13.653 sys 0m0.255s 00:17:13.653 14:16:18 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:13.653 14:16:18 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:13.653 ************************************ 00:17:13.653 END TEST bdev_verify 00:17:13.653 ************************************ 00:17:13.914 14:16:18 blockdev_raid5f -- bdev/blockdev.sh@777 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:13.914 14:16:18 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:13.914 14:16:18 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:13.914 14:16:18 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:13.914 ************************************ 00:17:13.914 START TEST bdev_verify_big_io 00:17:13.914 ************************************ 00:17:13.914 14:16:18 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:13.914 [2024-09-30 14:16:18.445074] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:17:13.914 [2024-09-30 14:16:18.445240] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101818 ] 00:17:14.174 [2024-09-30 14:16:18.583863] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:17:14.174 [2024-09-30 14:16:18.612311] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:14.174 [2024-09-30 14:16:18.663460] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:14.174 [2024-09-30 14:16:18.663624] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:14.434 Running I/O for 5 seconds... 00:17:19.566 633.00 IOPS, 39.56 MiB/s 761.00 IOPS, 47.56 MiB/s 803.67 IOPS, 50.23 MiB/s 825.00 IOPS, 51.56 MiB/s 838.00 IOPS, 52.38 MiB/s 00:17:19.566 Latency(us) 00:17:19.566 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:19.566 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:19.566 Verification LBA range: start 0x0 length 0x200 00:17:19.566 raid5f : 5.30 358.74 22.42 0.00 0.00 8794659.01 228.95 377304.20 00:17:19.566 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:19.566 Verification LBA range: start 0x200 length 0x200 00:17:19.566 raid5f : 5.20 464.37 29.02 0.00 0.00 6892874.93 216.43 298546.53 00:17:19.566 =================================================================================================================== 00:17:19.566 Total : 823.11 51.44 0.00 0.00 7731404.85 216.43 377304.20 00:17:19.826 00:17:19.826 real 0m6.064s 00:17:19.826 user 0m11.261s 00:17:19.826 sys 0m0.254s 00:17:19.826 14:16:24 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.826 14:16:24 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:19.826 ************************************ 00:17:19.826 END TEST bdev_verify_big_io 00:17:19.826 ************************************ 00:17:19.826 14:16:24 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:19.826 14:16:24 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:19.826 14:16:24 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:19.826 14:16:24 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:20.086 ************************************ 00:17:20.086 START TEST bdev_write_zeroes 00:17:20.086 ************************************ 00:17:20.086 14:16:24 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:20.086 [2024-09-30 14:16:24.576787] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:17:20.086 [2024-09-30 14:16:24.576916] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101900 ] 00:17:20.086 [2024-09-30 14:16:24.709005] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:17:20.086 [2024-09-30 14:16:24.738457] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:20.346 [2024-09-30 14:16:24.784896] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:20.346 Running I/O for 1 seconds... 00:17:21.727 30255.00 IOPS, 118.18 MiB/s 00:17:21.727 Latency(us) 00:17:21.727 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:21.727 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:21.727 raid5f : 1.01 30242.03 118.13 0.00 0.00 4221.72 1287.83 5695.05 00:17:21.727 =================================================================================================================== 00:17:21.727 Total : 30242.03 118.13 0.00 0.00 4221.72 1287.83 5695.05 00:17:21.727 00:17:21.727 real 0m1.755s 00:17:21.727 user 0m1.401s 00:17:21.727 sys 0m0.232s 00:17:21.727 14:16:26 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:21.727 14:16:26 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:21.727 ************************************ 00:17:21.727 END TEST bdev_write_zeroes 00:17:21.727 ************************************ 00:17:21.727 14:16:26 blockdev_raid5f -- bdev/blockdev.sh@781 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:21.727 14:16:26 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:21.727 14:16:26 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:21.727 14:16:26 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:21.727 ************************************ 00:17:21.727 START TEST bdev_json_nonenclosed 00:17:21.727 ************************************ 00:17:21.727 14:16:26 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:21.986 [2024-09-30 14:16:26.404317] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:17:21.987 [2024-09-30 14:16:26.404496] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101941 ] 00:17:21.987 [2024-09-30 14:16:26.534981] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:17:21.987 [2024-09-30 14:16:26.565929] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:21.987 [2024-09-30 14:16:26.618788] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:21.987 [2024-09-30 14:16:26.618892] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:21.987 [2024-09-30 14:16:26.618923] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:21.987 [2024-09-30 14:16:26.618933] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:22.246 00:17:22.246 real 0m0.415s 00:17:22.246 user 0m0.167s 00:17:22.246 sys 0m0.143s 00:17:22.246 14:16:26 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:22.246 14:16:26 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:22.246 ************************************ 00:17:22.247 END TEST bdev_json_nonenclosed 00:17:22.247 ************************************ 00:17:22.247 14:16:26 blockdev_raid5f -- bdev/blockdev.sh@784 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:22.247 14:16:26 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:22.247 14:16:26 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:22.247 14:16:26 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:22.247 ************************************ 00:17:22.247 START TEST bdev_json_nonarray 00:17:22.247 ************************************ 00:17:22.247 14:16:26 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:22.507 [2024-09-30 14:16:26.906152] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 24.11.0-rc0 initialization... 00:17:22.507 [2024-09-30 14:16:26.906276] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101962 ] 00:17:22.507 [2024-09-30 14:16:27.038001] pci_dpdk.c: 37:dpdk_pci_init: *NOTICE*: In-development DPDK 24.11.0-rc0 is used. There is no support for it in SPDK. Enabled only for validation. 00:17:22.507 [2024-09-30 14:16:27.066219] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:22.507 [2024-09-30 14:16:27.112646] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:22.507 [2024-09-30 14:16:27.112748] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:22.507 [2024-09-30 14:16:27.112770] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:22.507 [2024-09-30 14:16:27.112780] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:22.767 00:17:22.767 real 0m0.422s 00:17:22.767 user 0m0.181s 00:17:22.767 sys 0m0.135s 00:17:22.767 14:16:27 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:22.767 14:16:27 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:17:22.767 ************************************ 00:17:22.767 END TEST bdev_json_nonarray 00:17:22.767 ************************************ 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@786 -- # [[ raid5f == bdev ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@793 -- # [[ raid5f == gpt ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@797 -- # [[ raid5f == crypto_sw ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@809 -- # trap - SIGINT SIGTERM EXIT 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@810 -- # cleanup 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:17:22.767 14:16:27 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:17:22.767 00:17:22.767 real 0m35.136s 00:17:22.767 user 0m47.551s 00:17:22.767 sys 0m4.631s 00:17:22.767 14:16:27 blockdev_raid5f -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:22.767 14:16:27 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:22.767 ************************************ 00:17:22.767 END TEST blockdev_raid5f 00:17:22.767 ************************************ 00:17:22.767 14:16:27 -- spdk/autotest.sh@194 -- # uname -s 00:17:22.767 14:16:27 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:17:22.767 14:16:27 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:22.767 14:16:27 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:22.767 14:16:27 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:17:22.767 14:16:27 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:17:22.767 14:16:27 -- spdk/autotest.sh@256 -- # timing_exit lib 00:17:22.767 14:16:27 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:22.767 14:16:27 -- common/autotest_common.sh@10 -- # set +x 00:17:23.027 14:16:27 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@272 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:17:23.027 14:16:27 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:17:23.028 14:16:27 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:17:23.028 14:16:27 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:17:23.028 14:16:27 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:17:23.028 14:16:27 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:17:23.028 14:16:27 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:17:23.028 14:16:27 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:17:23.028 14:16:27 -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:23.028 14:16:27 -- common/autotest_common.sh@10 -- # set +x 00:17:23.028 14:16:27 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:17:23.028 14:16:27 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:17:23.028 14:16:27 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:17:23.028 14:16:27 -- common/autotest_common.sh@10 -- # set +x 00:17:25.568 INFO: APP EXITING 00:17:25.568 INFO: killing all VMs 00:17:25.568 INFO: killing vhost app 00:17:25.568 INFO: EXIT DONE 00:17:25.827 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:25.827 Waiting for block devices as requested 00:17:25.827 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:17:25.827 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:17:26.768 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:27.029 Cleaning 00:17:27.029 Removing: /var/run/dpdk/spdk0/config 00:17:27.029 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:17:27.029 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:17:27.029 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:17:27.029 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:17:27.029 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:17:27.029 Removing: /var/run/dpdk/spdk0/hugepage_info 00:17:27.029 Removing: /dev/shm/spdk_tgt_trace.pid70175 00:17:27.029 Removing: /var/run/dpdk/spdk0 00:17:27.029 Removing: /var/run/dpdk/spdk_pid100027 00:17:27.029 Removing: /var/run/dpdk/spdk_pid100340 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101010 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101272 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101308 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101339 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101574 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101736 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101818 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101900 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101941 00:17:27.029 Removing: /var/run/dpdk/spdk_pid101962 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70012 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70175 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70379 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70464 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70493 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70607 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70622 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70809 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70884 00:17:27.029 Removing: /var/run/dpdk/spdk_pid70969 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71068 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71144 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71189 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71220 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71296 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71397 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71823 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71871 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71922 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71934 00:17:27.029 Removing: /var/run/dpdk/spdk_pid71998 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72014 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72078 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72094 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72136 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72154 00:17:27.029 Removing: /var/run/dpdk/spdk_pid72196 00:17:27.289 Removing: /var/run/dpdk/spdk_pid72214 00:17:27.289 Removing: /var/run/dpdk/spdk_pid72352 00:17:27.289 Removing: /var/run/dpdk/spdk_pid72396 00:17:27.289 Removing: /var/run/dpdk/spdk_pid72474 00:17:27.289 Removing: /var/run/dpdk/spdk_pid73643 00:17:27.289 Removing: /var/run/dpdk/spdk_pid73838 00:17:27.289 Removing: /var/run/dpdk/spdk_pid73967 00:17:27.289 Removing: /var/run/dpdk/spdk_pid74576 00:17:27.289 Removing: /var/run/dpdk/spdk_pid74772 00:17:27.289 Removing: /var/run/dpdk/spdk_pid74901 00:17:27.289 Removing: /var/run/dpdk/spdk_pid75506 00:17:27.289 Removing: /var/run/dpdk/spdk_pid75819 00:17:27.289 Removing: /var/run/dpdk/spdk_pid75948 00:17:27.289 Removing: /var/run/dpdk/spdk_pid77284 00:17:27.289 Removing: /var/run/dpdk/spdk_pid77520 00:17:27.289 Removing: /var/run/dpdk/spdk_pid77654 00:17:27.289 Removing: /var/run/dpdk/spdk_pid78986 00:17:27.289 Removing: /var/run/dpdk/spdk_pid79228 00:17:27.289 Removing: /var/run/dpdk/spdk_pid79357 00:17:27.289 Removing: /var/run/dpdk/spdk_pid80698 00:17:27.289 Removing: /var/run/dpdk/spdk_pid81127 00:17:27.289 Removing: /var/run/dpdk/spdk_pid81260 00:17:27.289 Removing: /var/run/dpdk/spdk_pid82686 00:17:27.289 Removing: /var/run/dpdk/spdk_pid82934 00:17:27.289 Removing: /var/run/dpdk/spdk_pid83069 00:17:27.289 Removing: /var/run/dpdk/spdk_pid84493 00:17:27.289 Removing: /var/run/dpdk/spdk_pid84747 00:17:27.289 Removing: /var/run/dpdk/spdk_pid84876 00:17:27.289 Removing: /var/run/dpdk/spdk_pid86307 00:17:27.289 Removing: /var/run/dpdk/spdk_pid86777 00:17:27.289 Removing: /var/run/dpdk/spdk_pid86912 00:17:27.289 Removing: /var/run/dpdk/spdk_pid87039 00:17:27.289 Removing: /var/run/dpdk/spdk_pid87445 00:17:27.289 Removing: /var/run/dpdk/spdk_pid88162 00:17:27.289 Removing: /var/run/dpdk/spdk_pid88521 00:17:27.289 Removing: /var/run/dpdk/spdk_pid89220 00:17:27.289 Removing: /var/run/dpdk/spdk_pid89640 00:17:27.289 Removing: /var/run/dpdk/spdk_pid90376 00:17:27.289 Removing: /var/run/dpdk/spdk_pid90768 00:17:27.289 Removing: /var/run/dpdk/spdk_pid92677 00:17:27.289 Removing: /var/run/dpdk/spdk_pid93110 00:17:27.289 Removing: /var/run/dpdk/spdk_pid93533 00:17:27.289 Removing: /var/run/dpdk/spdk_pid95567 00:17:27.289 Removing: /var/run/dpdk/spdk_pid96036 00:17:27.289 Removing: /var/run/dpdk/spdk_pid96543 00:17:27.289 Removing: /var/run/dpdk/spdk_pid97572 00:17:27.289 Removing: /var/run/dpdk/spdk_pid97888 00:17:27.289 Removing: /var/run/dpdk/spdk_pid98800 00:17:27.289 Removing: /var/run/dpdk/spdk_pid99111 00:17:27.289 Clean 00:17:27.554 14:16:31 -- common/autotest_common.sh@1451 -- # return 0 00:17:27.554 14:16:31 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:17:27.554 14:16:31 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:27.554 14:16:31 -- common/autotest_common.sh@10 -- # set +x 00:17:27.554 14:16:32 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:17:27.554 14:16:32 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:27.554 14:16:32 -- common/autotest_common.sh@10 -- # set +x 00:17:27.554 14:16:32 -- spdk/autotest.sh@388 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:27.555 14:16:32 -- spdk/autotest.sh@390 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:17:27.555 14:16:32 -- spdk/autotest.sh@390 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:17:27.555 14:16:32 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:17:27.555 14:16:32 -- spdk/autotest.sh@394 -- # hostname 00:17:27.555 14:16:32 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:17:27.815 geninfo: WARNING: invalid characters removed from testname! 00:17:54.404 14:16:56 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:54.675 14:16:59 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:57.297 14:17:01 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:59.208 14:17:03 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:01.750 14:17:05 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:03.133 14:17:07 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:05.685 14:17:09 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:05.685 14:17:09 -- common/autotest_common.sh@1680 -- $ [[ y == y ]] 00:18:05.685 14:17:09 -- common/autotest_common.sh@1681 -- $ lcov --version 00:18:05.685 14:17:09 -- common/autotest_common.sh@1681 -- $ awk '{print $NF}' 00:18:05.685 14:17:10 -- common/autotest_common.sh@1681 -- $ lt 1.15 2 00:18:05.685 14:17:10 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:18:05.685 14:17:10 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:18:05.685 14:17:10 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:18:05.685 14:17:10 -- scripts/common.sh@336 -- $ IFS=.-: 00:18:05.685 14:17:10 -- scripts/common.sh@336 -- $ read -ra ver1 00:18:05.685 14:17:10 -- scripts/common.sh@337 -- $ IFS=.-: 00:18:05.685 14:17:10 -- scripts/common.sh@337 -- $ read -ra ver2 00:18:05.685 14:17:10 -- scripts/common.sh@338 -- $ local 'op=<' 00:18:05.685 14:17:10 -- scripts/common.sh@340 -- $ ver1_l=2 00:18:05.685 14:17:10 -- scripts/common.sh@341 -- $ ver2_l=1 00:18:05.685 14:17:10 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:18:05.685 14:17:10 -- scripts/common.sh@344 -- $ case "$op" in 00:18:05.685 14:17:10 -- scripts/common.sh@345 -- $ : 1 00:18:05.685 14:17:10 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:18:05.685 14:17:10 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:05.685 14:17:10 -- scripts/common.sh@365 -- $ decimal 1 00:18:05.685 14:17:10 -- scripts/common.sh@353 -- $ local d=1 00:18:05.685 14:17:10 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:18:05.685 14:17:10 -- scripts/common.sh@355 -- $ echo 1 00:18:05.685 14:17:10 -- scripts/common.sh@365 -- $ ver1[v]=1 00:18:05.685 14:17:10 -- scripts/common.sh@366 -- $ decimal 2 00:18:05.685 14:17:10 -- scripts/common.sh@353 -- $ local d=2 00:18:05.685 14:17:10 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:18:05.685 14:17:10 -- scripts/common.sh@355 -- $ echo 2 00:18:05.685 14:17:10 -- scripts/common.sh@366 -- $ ver2[v]=2 00:18:05.685 14:17:10 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:18:05.685 14:17:10 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:18:05.685 14:17:10 -- scripts/common.sh@368 -- $ return 0 00:18:05.685 14:17:10 -- common/autotest_common.sh@1682 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:05.685 14:17:10 -- common/autotest_common.sh@1694 -- $ export 'LCOV_OPTS= 00:18:05.685 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:05.685 --rc genhtml_branch_coverage=1 00:18:05.685 --rc genhtml_function_coverage=1 00:18:05.685 --rc genhtml_legend=1 00:18:05.685 --rc geninfo_all_blocks=1 00:18:05.685 --rc geninfo_unexecuted_blocks=1 00:18:05.685 00:18:05.685 ' 00:18:05.685 14:17:10 -- common/autotest_common.sh@1694 -- $ LCOV_OPTS=' 00:18:05.685 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:05.685 --rc genhtml_branch_coverage=1 00:18:05.685 --rc genhtml_function_coverage=1 00:18:05.685 --rc genhtml_legend=1 00:18:05.685 --rc geninfo_all_blocks=1 00:18:05.685 --rc geninfo_unexecuted_blocks=1 00:18:05.685 00:18:05.685 ' 00:18:05.685 14:17:10 -- common/autotest_common.sh@1695 -- $ export 'LCOV=lcov 00:18:05.685 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:05.685 --rc genhtml_branch_coverage=1 00:18:05.685 --rc genhtml_function_coverage=1 00:18:05.685 --rc genhtml_legend=1 00:18:05.685 --rc geninfo_all_blocks=1 00:18:05.685 --rc geninfo_unexecuted_blocks=1 00:18:05.685 00:18:05.685 ' 00:18:05.685 14:17:10 -- common/autotest_common.sh@1695 -- $ LCOV='lcov 00:18:05.685 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:05.685 --rc genhtml_branch_coverage=1 00:18:05.685 --rc genhtml_function_coverage=1 00:18:05.685 --rc genhtml_legend=1 00:18:05.685 --rc geninfo_all_blocks=1 00:18:05.685 --rc geninfo_unexecuted_blocks=1 00:18:05.685 00:18:05.685 ' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:18:05.685 14:17:10 -- scripts/common.sh@15 -- $ shopt -s extglob 00:18:05.685 14:17:10 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:18:05.685 14:17:10 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:05.685 14:17:10 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:05.685 14:17:10 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:05.685 14:17:10 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:05.685 14:17:10 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:05.685 14:17:10 -- paths/export.sh@5 -- $ export PATH 00:18:05.685 14:17:10 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:05.685 14:17:10 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:18:05.685 14:17:10 -- common/autobuild_common.sh@479 -- $ date +%s 00:18:05.685 14:17:10 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727705830.XXXXXX 00:18:05.685 14:17:10 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727705830.jhhAGV 00:18:05.685 14:17:10 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:18:05.685 14:17:10 -- common/autobuild_common.sh@485 -- $ '[' -n main ']' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:18:05.685 14:17:10 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@495 -- $ get_config_params 00:18:05.685 14:17:10 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:18:05.685 14:17:10 -- common/autotest_common.sh@10 -- $ set +x 00:18:05.685 14:17:10 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:18:05.685 14:17:10 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:18:05.685 14:17:10 -- pm/common@17 -- $ local monitor 00:18:05.685 14:17:10 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:05.686 14:17:10 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:05.686 14:17:10 -- pm/common@25 -- $ sleep 1 00:18:05.686 14:17:10 -- pm/common@21 -- $ date +%s 00:18:05.686 14:17:10 -- pm/common@21 -- $ date +%s 00:18:05.686 14:17:10 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727705830 00:18:05.686 14:17:10 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727705830 00:18:05.686 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727705830_collect-cpu-load.pm.log 00:18:05.686 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727705830_collect-vmstat.pm.log 00:18:06.627 14:17:11 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:18:06.627 14:17:11 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:18:06.627 14:17:11 -- spdk/autopackage.sh@14 -- $ timing_finish 00:18:06.627 14:17:11 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:06.627 14:17:11 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:06.627 14:17:11 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:06.628 14:17:11 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:18:06.628 14:17:11 -- pm/common@29 -- $ signal_monitor_resources TERM 00:18:06.628 14:17:11 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:18:06.628 14:17:11 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:06.628 14:17:11 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:18:06.628 14:17:11 -- pm/common@44 -- $ pid=103497 00:18:06.628 14:17:11 -- pm/common@50 -- $ kill -TERM 103497 00:18:06.628 14:17:11 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:06.628 14:17:11 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:18:06.628 14:17:11 -- pm/common@44 -- $ pid=103499 00:18:06.628 14:17:11 -- pm/common@50 -- $ kill -TERM 103499 00:18:06.628 + [[ -n 6169 ]] 00:18:06.628 + sudo kill 6169 00:18:06.638 [Pipeline] } 00:18:06.655 [Pipeline] // timeout 00:18:06.660 [Pipeline] } 00:18:06.675 [Pipeline] // stage 00:18:06.681 [Pipeline] } 00:18:06.707 [Pipeline] // catchError 00:18:06.717 [Pipeline] stage 00:18:06.719 [Pipeline] { (Stop VM) 00:18:06.732 [Pipeline] sh 00:18:07.018 + vagrant halt 00:18:09.556 ==> default: Halting domain... 00:18:17.724 [Pipeline] sh 00:18:18.007 + vagrant destroy -f 00:18:20.549 ==> default: Removing domain... 00:18:20.563 [Pipeline] sh 00:18:20.850 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:20.861 [Pipeline] } 00:18:20.878 [Pipeline] // stage 00:18:20.884 [Pipeline] } 00:18:20.902 [Pipeline] // dir 00:18:20.908 [Pipeline] } 00:18:20.926 [Pipeline] // wrap 00:18:20.933 [Pipeline] } 00:18:20.948 [Pipeline] // catchError 00:18:20.958 [Pipeline] stage 00:18:20.960 [Pipeline] { (Epilogue) 00:18:20.974 [Pipeline] sh 00:18:21.259 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:18:25.470 [Pipeline] catchError 00:18:25.472 [Pipeline] { 00:18:25.486 [Pipeline] sh 00:18:25.771 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:18:25.772 Artifacts sizes are good 00:18:25.782 [Pipeline] } 00:18:25.797 [Pipeline] // catchError 00:18:25.808 [Pipeline] archiveArtifacts 00:18:25.815 Archiving artifacts 00:18:25.961 [Pipeline] cleanWs 00:18:25.974 [WS-CLEANUP] Deleting project workspace... 00:18:25.974 [WS-CLEANUP] Deferred wipeout is used... 00:18:25.992 [WS-CLEANUP] done 00:18:25.994 [Pipeline] } 00:18:26.009 [Pipeline] // stage 00:18:26.014 [Pipeline] } 00:18:26.027 [Pipeline] // node 00:18:26.033 [Pipeline] End of Pipeline 00:18:26.078 Finished: SUCCESS